Mar 21 04:47:25 crc systemd[1]: Starting Kubernetes Kubelet... Mar 21 04:47:25 crc restorecon[4688]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 04:47:25 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 04:47:26 crc restorecon[4688]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 04:47:26 crc restorecon[4688]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 21 04:47:27 crc kubenswrapper[4775]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 21 04:47:27 crc kubenswrapper[4775]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 21 04:47:27 crc kubenswrapper[4775]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 21 04:47:27 crc kubenswrapper[4775]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 21 04:47:27 crc kubenswrapper[4775]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 21 04:47:27 crc kubenswrapper[4775]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.424970 4775 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431082 4775 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431105 4775 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431127 4775 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431133 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431137 4775 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431142 4775 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431146 4775 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431151 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431155 4775 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431159 4775 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431163 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431167 4775 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431171 4775 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431174 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431179 4775 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431195 4775 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431201 4775 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431206 4775 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431212 4775 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431217 4775 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431223 4775 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431229 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431235 4775 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431240 4775 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431244 4775 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431248 4775 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431252 4775 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431257 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431260 4775 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431264 4775 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431269 4775 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431274 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431278 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431284 4775 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431289 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431293 4775 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431299 4775 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431305 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431310 4775 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431314 4775 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431318 4775 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431323 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431328 4775 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431332 4775 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431337 4775 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431341 4775 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431345 4775 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431350 4775 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431355 4775 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431359 4775 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431363 4775 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431367 4775 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431371 4775 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431378 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431382 4775 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431386 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431389 4775 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431393 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431397 4775 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431401 4775 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431406 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431410 4775 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431415 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431421 4775 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431426 4775 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431431 4775 feature_gate.go:330] unrecognized feature gate: Example Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431436 4775 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431441 4775 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431445 4775 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431450 4775 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.431456 4775 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432104 4775 flags.go:64] FLAG: --address="0.0.0.0" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432135 4775 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432145 4775 flags.go:64] FLAG: --anonymous-auth="true" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432152 4775 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432159 4775 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432164 4775 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432172 4775 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432178 4775 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432183 4775 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432187 4775 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432192 4775 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432197 4775 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432202 4775 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432206 4775 flags.go:64] FLAG: --cgroup-root="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432211 4775 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432216 4775 flags.go:64] FLAG: --client-ca-file="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432221 4775 flags.go:64] FLAG: --cloud-config="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432225 4775 flags.go:64] FLAG: --cloud-provider="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432229 4775 flags.go:64] FLAG: --cluster-dns="[]" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432236 4775 flags.go:64] FLAG: --cluster-domain="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432241 4775 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432246 4775 flags.go:64] FLAG: --config-dir="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432250 4775 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432256 4775 flags.go:64] FLAG: --container-log-max-files="5" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432263 4775 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432268 4775 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432274 4775 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432279 4775 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432284 4775 flags.go:64] FLAG: --contention-profiling="false" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432289 4775 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432294 4775 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432300 4775 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432305 4775 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432312 4775 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432318 4775 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432324 4775 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432330 4775 flags.go:64] FLAG: --enable-load-reader="false" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432337 4775 flags.go:64] FLAG: --enable-server="true" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432342 4775 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432350 4775 flags.go:64] FLAG: --event-burst="100" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432356 4775 flags.go:64] FLAG: --event-qps="50" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432363 4775 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432368 4775 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432374 4775 flags.go:64] FLAG: --eviction-hard="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432381 4775 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432386 4775 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432392 4775 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432398 4775 flags.go:64] FLAG: --eviction-soft="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432403 4775 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432408 4775 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432413 4775 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432418 4775 flags.go:64] FLAG: --experimental-mounter-path="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432423 4775 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432428 4775 flags.go:64] FLAG: --fail-swap-on="true" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432433 4775 flags.go:64] FLAG: --feature-gates="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432440 4775 flags.go:64] FLAG: --file-check-frequency="20s" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432445 4775 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432450 4775 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432455 4775 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432461 4775 flags.go:64] FLAG: --healthz-port="10248" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432469 4775 flags.go:64] FLAG: --help="false" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432474 4775 flags.go:64] FLAG: --hostname-override="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432480 4775 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432485 4775 flags.go:64] FLAG: --http-check-frequency="20s" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432490 4775 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432495 4775 flags.go:64] FLAG: --image-credential-provider-config="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432500 4775 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432505 4775 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432510 4775 flags.go:64] FLAG: --image-service-endpoint="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432515 4775 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432520 4775 flags.go:64] FLAG: --kube-api-burst="100" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432525 4775 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432530 4775 flags.go:64] FLAG: --kube-api-qps="50" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432535 4775 flags.go:64] FLAG: --kube-reserved="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432540 4775 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432545 4775 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432550 4775 flags.go:64] FLAG: --kubelet-cgroups="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432555 4775 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432560 4775 flags.go:64] FLAG: --lock-file="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432565 4775 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432570 4775 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432575 4775 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432583 4775 flags.go:64] FLAG: --log-json-split-stream="false" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432588 4775 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432592 4775 flags.go:64] FLAG: --log-text-split-stream="false" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432597 4775 flags.go:64] FLAG: --logging-format="text" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432602 4775 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432608 4775 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432613 4775 flags.go:64] FLAG: --manifest-url="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432618 4775 flags.go:64] FLAG: --manifest-url-header="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432625 4775 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432630 4775 flags.go:64] FLAG: --max-open-files="1000000" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432637 4775 flags.go:64] FLAG: --max-pods="110" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432643 4775 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432648 4775 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432654 4775 flags.go:64] FLAG: --memory-manager-policy="None" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432658 4775 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432664 4775 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432668 4775 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432673 4775 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432686 4775 flags.go:64] FLAG: --node-status-max-images="50" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432691 4775 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432696 4775 flags.go:64] FLAG: --oom-score-adj="-999" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432701 4775 flags.go:64] FLAG: --pod-cidr="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432707 4775 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432716 4775 flags.go:64] FLAG: --pod-manifest-path="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432731 4775 flags.go:64] FLAG: --pod-max-pids="-1" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432736 4775 flags.go:64] FLAG: --pods-per-core="0" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432741 4775 flags.go:64] FLAG: --port="10250" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432746 4775 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432751 4775 flags.go:64] FLAG: --provider-id="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432755 4775 flags.go:64] FLAG: --qos-reserved="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432760 4775 flags.go:64] FLAG: --read-only-port="10255" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432765 4775 flags.go:64] FLAG: --register-node="true" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432771 4775 flags.go:64] FLAG: --register-schedulable="true" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432777 4775 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432787 4775 flags.go:64] FLAG: --registry-burst="10" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432793 4775 flags.go:64] FLAG: --registry-qps="5" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432799 4775 flags.go:64] FLAG: --reserved-cpus="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432804 4775 flags.go:64] FLAG: --reserved-memory="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432811 4775 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432816 4775 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432822 4775 flags.go:64] FLAG: --rotate-certificates="false" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432827 4775 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432832 4775 flags.go:64] FLAG: --runonce="false" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432837 4775 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432842 4775 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432847 4775 flags.go:64] FLAG: --seccomp-default="false" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432852 4775 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432857 4775 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432862 4775 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432868 4775 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432873 4775 flags.go:64] FLAG: --storage-driver-password="root" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432878 4775 flags.go:64] FLAG: --storage-driver-secure="false" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432883 4775 flags.go:64] FLAG: --storage-driver-table="stats" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432889 4775 flags.go:64] FLAG: --storage-driver-user="root" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432894 4775 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432899 4775 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432909 4775 flags.go:64] FLAG: --system-cgroups="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432914 4775 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432923 4775 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432928 4775 flags.go:64] FLAG: --tls-cert-file="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432933 4775 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432939 4775 flags.go:64] FLAG: --tls-min-version="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432944 4775 flags.go:64] FLAG: --tls-private-key-file="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432949 4775 flags.go:64] FLAG: --topology-manager-policy="none" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432954 4775 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432959 4775 flags.go:64] FLAG: --topology-manager-scope="container" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432964 4775 flags.go:64] FLAG: --v="2" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432971 4775 flags.go:64] FLAG: --version="false" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432979 4775 flags.go:64] FLAG: --vmodule="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432986 4775 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.432991 4775 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433108 4775 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433139 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433145 4775 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433150 4775 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433156 4775 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433161 4775 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433166 4775 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433171 4775 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433175 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433180 4775 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433184 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433189 4775 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433193 4775 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433197 4775 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433201 4775 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433206 4775 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433210 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433222 4775 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433228 4775 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433234 4775 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433239 4775 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433244 4775 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433249 4775 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433255 4775 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433259 4775 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433264 4775 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433269 4775 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433274 4775 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433279 4775 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433284 4775 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433291 4775 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433297 4775 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433302 4775 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433307 4775 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433314 4775 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433322 4775 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433327 4775 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433332 4775 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433338 4775 feature_gate.go:330] unrecognized feature gate: Example Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433343 4775 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433348 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433352 4775 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433356 4775 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433361 4775 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433365 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433369 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433373 4775 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433393 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433398 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433406 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433411 4775 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433415 4775 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433420 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433424 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433428 4775 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433433 4775 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433437 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433442 4775 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433446 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433452 4775 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433456 4775 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433461 4775 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433465 4775 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433470 4775 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433474 4775 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433478 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433482 4775 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433486 4775 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433491 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433495 4775 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.433499 4775 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.433507 4775 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.445079 4775 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.445144 4775 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445224 4775 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445234 4775 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445239 4775 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445245 4775 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445249 4775 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445253 4775 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445256 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445260 4775 feature_gate.go:330] unrecognized feature gate: Example Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445264 4775 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445268 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445271 4775 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445275 4775 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445278 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445281 4775 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445285 4775 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445289 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445292 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445297 4775 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445303 4775 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445308 4775 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445311 4775 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445316 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445319 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445323 4775 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445327 4775 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445331 4775 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445336 4775 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445343 4775 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445347 4775 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445352 4775 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445356 4775 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445362 4775 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445368 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445373 4775 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445386 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445392 4775 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445396 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445401 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445405 4775 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445410 4775 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445414 4775 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445418 4775 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445423 4775 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445429 4775 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445438 4775 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445442 4775 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445447 4775 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445451 4775 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445456 4775 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445460 4775 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445465 4775 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445471 4775 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445476 4775 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445483 4775 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445488 4775 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445493 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445498 4775 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445502 4775 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445505 4775 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445508 4775 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445512 4775 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445516 4775 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445519 4775 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445523 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445526 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445530 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445533 4775 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445537 4775 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445540 4775 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445543 4775 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445555 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.445564 4775 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445748 4775 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445765 4775 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445772 4775 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445779 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445783 4775 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445787 4775 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445791 4775 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445794 4775 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445798 4775 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445807 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445811 4775 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445814 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445817 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445823 4775 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445828 4775 feature_gate.go:330] unrecognized feature gate: Example Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445831 4775 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445835 4775 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445838 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445841 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445845 4775 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445849 4775 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445852 4775 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445855 4775 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445859 4775 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445862 4775 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445867 4775 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445872 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445875 4775 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445879 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445882 4775 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445886 4775 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445889 4775 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445893 4775 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445897 4775 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445906 4775 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445910 4775 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445913 4775 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445917 4775 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445921 4775 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445925 4775 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445929 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445932 4775 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445968 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445972 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445975 4775 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445979 4775 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445984 4775 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445988 4775 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445991 4775 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445995 4775 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.445998 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.446003 4775 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.446007 4775 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.446011 4775 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.446014 4775 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.446018 4775 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.446021 4775 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.446025 4775 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.446028 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.446032 4775 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.446035 4775 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.446038 4775 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.446042 4775 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.446045 4775 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.446049 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.446052 4775 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.446055 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.446059 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.446062 4775 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.446065 4775 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.446076 4775 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.446082 4775 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.448669 4775 server.go:940] "Client rotation is on, will bootstrap in background" Mar 21 04:47:27 crc kubenswrapper[4775]: E0321 04:47:27.452838 4775 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.457266 4775 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.457524 4775 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.459409 4775 server.go:997] "Starting client certificate rotation" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.459445 4775 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.460593 4775 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.493205 4775 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 21 04:47:27 crc kubenswrapper[4775]: E0321 04:47:27.496659 4775 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.497090 4775 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.514187 4775 log.go:25] "Validated CRI v1 runtime API" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.546058 4775 log.go:25] "Validated CRI v1 image API" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.549275 4775 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.555706 4775 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-21-04-43-12-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.555735 4775 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.575221 4775 manager.go:217] Machine: {Timestamp:2026-03-21 04:47:27.569507451 +0000 UTC m=+0.545971095 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654132736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:b8f8f0e2-c78b-43d3-976e-2a86ca08a185 BootID:82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07 Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827068416 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730829824 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:e7:06:8a Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:e7:06:8a Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:f6:e6:43 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:8a:d6:aa Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:ad:52:11 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:24:cd:92 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:da:5b:8e:5c:8e:2d Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:be:5b:74:dd:99:a7 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654132736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.575574 4775 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.575899 4775 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.578684 4775 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.578916 4775 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.578952 4775 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.579257 4775 topology_manager.go:138] "Creating topology manager with none policy" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.579275 4775 container_manager_linux.go:303] "Creating device plugin manager" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.579749 4775 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.579787 4775 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.580050 4775 state_mem.go:36] "Initialized new in-memory state store" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.580438 4775 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.584823 4775 kubelet.go:418] "Attempting to sync node with API server" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.584871 4775 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.584908 4775 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.584926 4775 kubelet.go:324] "Adding apiserver pod source" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.584949 4775 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.589300 4775 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.589735 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Mar 21 04:47:27 crc kubenswrapper[4775]: E0321 04:47:27.589813 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.589765 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Mar 21 04:47:27 crc kubenswrapper[4775]: E0321 04:47:27.589894 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.590470 4775 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.593528 4775 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.595167 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.595188 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.595195 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.595202 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.595213 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.595219 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.595226 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.595236 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.595245 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.595255 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.595272 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.595280 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.597253 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.597691 4775 server.go:1280] "Started kubelet" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.598915 4775 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.599012 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.599036 4775 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 21 04:47:27 crc systemd[1]: Started Kubernetes Kubelet. Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.599680 4775 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.600906 4775 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.600958 4775 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.601204 4775 server.go:460] "Adding debug handlers to kubelet server" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.601598 4775 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.601651 4775 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 21 04:47:27 crc kubenswrapper[4775]: E0321 04:47:27.601531 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.601869 4775 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 21 04:47:27 crc kubenswrapper[4775]: E0321 04:47:27.602978 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="200ms" Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.603187 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Mar 21 04:47:27 crc kubenswrapper[4775]: E0321 04:47:27.603248 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.612970 4775 factory.go:55] Registering systemd factory Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.613024 4775 factory.go:221] Registration of the systemd container factory successfully Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.613622 4775 factory.go:153] Registering CRI-O factory Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.613651 4775 factory.go:221] Registration of the crio container factory successfully Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.613798 4775 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.613851 4775 factory.go:103] Registering Raw factory Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.613888 4775 manager.go:1196] Started watching for new ooms in manager Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.617009 4775 manager.go:319] Starting recovery of all containers Mar 21 04:47:27 crc kubenswrapper[4775]: E0321 04:47:27.616971 4775 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.200:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189ec1da8a498a5d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.597660765 +0000 UTC m=+0.574124389,LastTimestamp:2026-03-21 04:47:27.597660765 +0000 UTC m=+0.574124389,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619316 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619395 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619412 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619426 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619440 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619483 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619497 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619513 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619531 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619546 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619561 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619574 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619588 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619604 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619616 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619630 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619643 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619658 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619672 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619686 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619704 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619721 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619737 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619773 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619786 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619799 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619815 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619831 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619845 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619860 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619873 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619917 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619932 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619945 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619958 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619975 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.619989 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620021 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620035 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620051 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620065 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620082 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620135 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620151 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620165 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620181 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620195 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620209 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620223 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620237 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620252 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620266 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620288 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620303 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620319 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620335 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620350 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620365 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620379 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620393 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620410 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620424 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620438 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620452 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620466 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620479 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620495 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620508 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620522 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620535 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620550 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620570 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620584 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620599 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620614 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620628 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620641 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620658 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620672 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620686 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620700 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620715 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620729 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620744 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620760 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620776 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.620792 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.622902 4775 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.622958 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.622982 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623005 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623027 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623050 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623072 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623096 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623188 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623213 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623233 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623253 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623273 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623292 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623329 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623355 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623376 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623397 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623479 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623503 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623524 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623547 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623568 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623588 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623608 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623632 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623654 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623675 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623696 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623716 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623734 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623758 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623779 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623799 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623822 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623841 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623861 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623881 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623903 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623922 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623944 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623964 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.623983 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624003 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624024 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624044 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624067 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624088 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624108 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624152 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624172 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624191 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624212 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624231 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624250 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624268 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624287 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624307 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624326 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624346 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624366 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624385 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624402 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624421 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624442 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624461 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624479 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624499 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624517 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624537 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624557 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624577 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624596 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624617 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624636 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624654 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624673 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624692 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624712 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624733 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624753 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624771 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624795 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624814 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624833 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624854 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624873 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624893 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624915 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624935 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624953 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624973 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.624992 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625011 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625030 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625052 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625072 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625143 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625164 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625187 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625206 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625225 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625244 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625262 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625281 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625301 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625321 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625341 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625359 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625378 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625401 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625422 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625441 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625459 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625481 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625500 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625518 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625538 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625556 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625576 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625597 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625615 4775 reconstruct.go:97] "Volume reconstruction finished" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.625628 4775 reconciler.go:26] "Reconciler: start to sync state" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.639645 4775 manager.go:324] Recovery completed Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.650253 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.653425 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.653476 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.653491 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.654136 4775 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.654157 4775 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.654232 4775 state_mem.go:36] "Initialized new in-memory state store" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.658543 4775 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.660012 4775 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.660063 4775 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.660097 4775 kubelet.go:2335] "Starting kubelet main sync loop" Mar 21 04:47:27 crc kubenswrapper[4775]: E0321 04:47:27.660168 4775 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 21 04:47:27 crc kubenswrapper[4775]: W0321 04:47:27.661650 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Mar 21 04:47:27 crc kubenswrapper[4775]: E0321 04:47:27.661716 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.676003 4775 policy_none.go:49] "None policy: Start" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.677187 4775 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.677233 4775 state_mem.go:35] "Initializing new in-memory state store" Mar 21 04:47:27 crc kubenswrapper[4775]: E0321 04:47:27.702793 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.726006 4775 manager.go:334] "Starting Device Plugin manager" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.726082 4775 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.726095 4775 server.go:79] "Starting device plugin registration server" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.726577 4775 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.726593 4775 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.728020 4775 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.728168 4775 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.728179 4775 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 21 04:47:27 crc kubenswrapper[4775]: E0321 04:47:27.732593 4775 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.760953 4775 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.761156 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.762860 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.762912 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.762926 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.763093 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.763951 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.763980 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.763991 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.765396 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.765440 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.765473 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.765565 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.765616 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.766309 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.766341 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.766354 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.766486 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.766707 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.766765 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.767274 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.767310 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.767323 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.767461 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.767491 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.767503 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.767672 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.767687 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.767691 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.767753 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.767783 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.767831 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.768393 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.768421 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.768434 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.768398 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.768581 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.768594 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.768630 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.768662 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.768900 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.768935 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.768947 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.770747 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.770770 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.770782 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:27 crc kubenswrapper[4775]: E0321 04:47:27.803766 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="400ms" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.827435 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.828734 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.828789 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.828800 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.828941 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.829545 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.829589 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.829617 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.829639 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.829660 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.829683 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: E0321 04:47:27.829736 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.200:6443: connect: connection refused" node="crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.829745 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.829828 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.829859 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.829881 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.829905 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.829926 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.829976 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.830048 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.830155 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931480 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931544 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931576 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931602 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931631 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931657 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931689 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931723 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931755 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931756 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931785 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931810 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931869 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931749 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931875 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931939 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931907 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931814 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.932034 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.932082 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.932111 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931866 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.932182 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.932230 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.932304 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931940 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.932182 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.932235 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931904 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:47:27 crc kubenswrapper[4775]: I0321 04:47:27.931938 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.030005 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.031876 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.031995 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.032102 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.032229 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 04:47:28 crc kubenswrapper[4775]: E0321 04:47:28.033003 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.200:6443: connect: connection refused" node="crc" Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.097744 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.106772 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.123428 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.147518 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.153070 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:47:28 crc kubenswrapper[4775]: W0321 04:47:28.163080 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-c5c1d5c259f2beab5080951416bfd5289ddef57bb979d944747ea4b3e7081287 WatchSource:0}: Error finding container c5c1d5c259f2beab5080951416bfd5289ddef57bb979d944747ea4b3e7081287: Status 404 returned error can't find the container with id c5c1d5c259f2beab5080951416bfd5289ddef57bb979d944747ea4b3e7081287 Mar 21 04:47:28 crc kubenswrapper[4775]: W0321 04:47:28.164228 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-0650b8396f80992e97f0325700c41c51f8284160620600d88f75773714af8b27 WatchSource:0}: Error finding container 0650b8396f80992e97f0325700c41c51f8284160620600d88f75773714af8b27: Status 404 returned error can't find the container with id 0650b8396f80992e97f0325700c41c51f8284160620600d88f75773714af8b27 Mar 21 04:47:28 crc kubenswrapper[4775]: W0321 04:47:28.170771 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-1ca29dbcf983f375a8a59480fdd65ddd29e56908139fcdd17346f8cc6cf01486 WatchSource:0}: Error finding container 1ca29dbcf983f375a8a59480fdd65ddd29e56908139fcdd17346f8cc6cf01486: Status 404 returned error can't find the container with id 1ca29dbcf983f375a8a59480fdd65ddd29e56908139fcdd17346f8cc6cf01486 Mar 21 04:47:28 crc kubenswrapper[4775]: W0321 04:47:28.177936 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-cd4a716363e7c11843b2e9634ff5e9e4bf620eb979fbe6e3cfb9e894c3c338a9 WatchSource:0}: Error finding container cd4a716363e7c11843b2e9634ff5e9e4bf620eb979fbe6e3cfb9e894c3c338a9: Status 404 returned error can't find the container with id cd4a716363e7c11843b2e9634ff5e9e4bf620eb979fbe6e3cfb9e894c3c338a9 Mar 21 04:47:28 crc kubenswrapper[4775]: W0321 04:47:28.191162 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-47a8bc8f4176dcd8f7bcc28e3dfececb6304138c76584ddd9b3996ffbac6af07 WatchSource:0}: Error finding container 47a8bc8f4176dcd8f7bcc28e3dfececb6304138c76584ddd9b3996ffbac6af07: Status 404 returned error can't find the container with id 47a8bc8f4176dcd8f7bcc28e3dfececb6304138c76584ddd9b3996ffbac6af07 Mar 21 04:47:28 crc kubenswrapper[4775]: E0321 04:47:28.204861 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="800ms" Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.433781 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.435443 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.435478 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.435489 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.435514 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 04:47:28 crc kubenswrapper[4775]: E0321 04:47:28.435939 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.200:6443: connect: connection refused" node="crc" Mar 21 04:47:28 crc kubenswrapper[4775]: W0321 04:47:28.473192 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Mar 21 04:47:28 crc kubenswrapper[4775]: E0321 04:47:28.473279 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.600646 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Mar 21 04:47:28 crc kubenswrapper[4775]: W0321 04:47:28.642890 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Mar 21 04:47:28 crc kubenswrapper[4775]: E0321 04:47:28.642981 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.664890 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"0650b8396f80992e97f0325700c41c51f8284160620600d88f75773714af8b27"} Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.665790 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"47a8bc8f4176dcd8f7bcc28e3dfececb6304138c76584ddd9b3996ffbac6af07"} Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.666726 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cd4a716363e7c11843b2e9634ff5e9e4bf620eb979fbe6e3cfb9e894c3c338a9"} Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.667762 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1ca29dbcf983f375a8a59480fdd65ddd29e56908139fcdd17346f8cc6cf01486"} Mar 21 04:47:28 crc kubenswrapper[4775]: I0321 04:47:28.668582 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c5c1d5c259f2beab5080951416bfd5289ddef57bb979d944747ea4b3e7081287"} Mar 21 04:47:28 crc kubenswrapper[4775]: W0321 04:47:28.674421 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Mar 21 04:47:28 crc kubenswrapper[4775]: E0321 04:47:28.674501 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Mar 21 04:47:28 crc kubenswrapper[4775]: W0321 04:47:28.897462 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Mar 21 04:47:28 crc kubenswrapper[4775]: E0321 04:47:28.897841 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Mar 21 04:47:29 crc kubenswrapper[4775]: E0321 04:47:29.006226 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="1.6s" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.236405 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.237739 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.237774 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.237783 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.237804 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 04:47:29 crc kubenswrapper[4775]: E0321 04:47:29.238300 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.200:6443: connect: connection refused" node="crc" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.600537 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.601599 4775 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 21 04:47:29 crc kubenswrapper[4775]: E0321 04:47:29.602514 4775 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.672473 4775 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec" exitCode=0 Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.672571 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec"} Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.672692 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.673664 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.673691 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.673700 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.674591 4775 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057" exitCode=0 Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.674629 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057"} Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.674797 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.675574 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.675843 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.675861 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.675869 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.676390 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.676407 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.676416 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.677233 4775 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="99787fe5e44de1b5adaa5d1b49dcb6db27a4176d232dc67119b5f633bfa4c948" exitCode=0 Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.677301 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"99787fe5e44de1b5adaa5d1b49dcb6db27a4176d232dc67119b5f633bfa4c948"} Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.677355 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.678927 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.678962 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.678977 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.680338 4775 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="375572b0f0f7a6917d62ef1138d8651eb2d6118d420fc647d1d24c29e793f17d" exitCode=0 Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.680503 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.680614 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"375572b0f0f7a6917d62ef1138d8651eb2d6118d420fc647d1d24c29e793f17d"} Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.681854 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.681887 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.681900 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.686207 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9299f9db8cbdd2dbd04e47bd1eba4f95724ec3dd04ec5083564dc0c38a960457"} Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.686241 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8cb44c7daffbeddbb39314ec11bd3235996e19b520875337dcac1ffd14f35953"} Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.686261 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"851dab44c58218d203e180465f8a216bfeec8271e78102856d11545520feabb4"} Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.686273 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"75cf7f90f0971bce2e51690954080e297523b92a5588c8effa75832879f0dd6b"} Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.686345 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.687070 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.687098 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:29 crc kubenswrapper[4775]: I0321 04:47:29.687109 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:30 crc kubenswrapper[4775]: W0321 04:47:30.301853 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Mar 21 04:47:30 crc kubenswrapper[4775]: E0321 04:47:30.301963 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Mar 21 04:47:30 crc kubenswrapper[4775]: W0321 04:47:30.305383 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Mar 21 04:47:30 crc kubenswrapper[4775]: E0321 04:47:30.305424 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Mar 21 04:47:30 crc kubenswrapper[4775]: W0321 04:47:30.439566 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Mar 21 04:47:30 crc kubenswrapper[4775]: E0321 04:47:30.439652 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.599737 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Mar 21 04:47:30 crc kubenswrapper[4775]: E0321 04:47:30.607706 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="3.2s" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.689283 4775 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539" exitCode=0 Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.689340 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539"} Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.689364 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.690962 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.691002 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.691013 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.693536 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5a3e2381fa82bd51ba8be325e0a83a387ebd790d6a91283d604ab46f9ae7d45e"} Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.693561 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"71042d514c08b225f10915e55a0ef253281a017b05412ca810a5fbbfee272bd5"} Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.693574 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2f04df41528bb781c23382f46200bd21b1d0e6064f535d4d47cde4cabedef8c0"} Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.693606 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.694712 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.694749 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.694759 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.695538 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"0cf88511498eb0f6182819ecb78616730fa9bfbe187057ea891f3ba73e550287"} Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.695562 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.696217 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.696246 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.696256 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.699855 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948"} Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.699889 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb"} Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.699893 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.699904 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc"} Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.699914 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70"} Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.700780 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.700803 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.700811 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.838998 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.844239 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.844305 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.844316 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:30 crc kubenswrapper[4775]: I0321 04:47:30.844352 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 04:47:30 crc kubenswrapper[4775]: E0321 04:47:30.845039 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.200:6443: connect: connection refused" node="crc" Mar 21 04:47:30 crc kubenswrapper[4775]: W0321 04:47:30.848932 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Mar 21 04:47:30 crc kubenswrapper[4775]: E0321 04:47:30.849050 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Mar 21 04:47:31 crc kubenswrapper[4775]: E0321 04:47:31.370094 4775 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.200:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189ec1da8a498a5d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.597660765 +0000 UTC m=+0.574124389,LastTimestamp:2026-03-21 04:47:27.597660765 +0000 UTC m=+0.574124389,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:31 crc kubenswrapper[4775]: I0321 04:47:31.710036 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c38b09928e8a15368f55622732324802c7af7bc3d13c7cf2d1d00b4ac43fef94"} Mar 21 04:47:31 crc kubenswrapper[4775]: I0321 04:47:31.710175 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:31 crc kubenswrapper[4775]: I0321 04:47:31.712852 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:31 crc kubenswrapper[4775]: I0321 04:47:31.712901 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:31 crc kubenswrapper[4775]: I0321 04:47:31.712923 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:31 crc kubenswrapper[4775]: I0321 04:47:31.716542 4775 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645" exitCode=0 Mar 21 04:47:31 crc kubenswrapper[4775]: I0321 04:47:31.716658 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:31 crc kubenswrapper[4775]: I0321 04:47:31.717388 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:31 crc kubenswrapper[4775]: I0321 04:47:31.717955 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645"} Mar 21 04:47:31 crc kubenswrapper[4775]: I0321 04:47:31.718055 4775 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 21 04:47:31 crc kubenswrapper[4775]: I0321 04:47:31.718087 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:31 crc kubenswrapper[4775]: I0321 04:47:31.719271 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:31 crc kubenswrapper[4775]: I0321 04:47:31.719309 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:31 crc kubenswrapper[4775]: I0321 04:47:31.719326 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:31 crc kubenswrapper[4775]: I0321 04:47:31.720069 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:31 crc kubenswrapper[4775]: I0321 04:47:31.720107 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:31 crc kubenswrapper[4775]: I0321 04:47:31.720151 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:31 crc kubenswrapper[4775]: I0321 04:47:31.720817 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:31 crc kubenswrapper[4775]: I0321 04:47:31.720883 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:31 crc kubenswrapper[4775]: I0321 04:47:31.720906 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:32 crc kubenswrapper[4775]: I0321 04:47:32.394111 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:47:32 crc kubenswrapper[4775]: I0321 04:47:32.394288 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:32 crc kubenswrapper[4775]: I0321 04:47:32.395315 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:32 crc kubenswrapper[4775]: I0321 04:47:32.395363 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:32 crc kubenswrapper[4775]: I0321 04:47:32.395378 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:32 crc kubenswrapper[4775]: I0321 04:47:32.722107 4775 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 21 04:47:32 crc kubenswrapper[4775]: I0321 04:47:32.722177 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:32 crc kubenswrapper[4775]: I0321 04:47:32.722564 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939"} Mar 21 04:47:32 crc kubenswrapper[4775]: I0321 04:47:32.722592 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987"} Mar 21 04:47:32 crc kubenswrapper[4775]: I0321 04:47:32.722603 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846"} Mar 21 04:47:32 crc kubenswrapper[4775]: I0321 04:47:32.722610 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4"} Mar 21 04:47:32 crc kubenswrapper[4775]: I0321 04:47:32.722893 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:32 crc kubenswrapper[4775]: I0321 04:47:32.722929 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:32 crc kubenswrapper[4775]: I0321 04:47:32.722939 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:33 crc kubenswrapper[4775]: I0321 04:47:33.132808 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:47:33 crc kubenswrapper[4775]: I0321 04:47:33.276444 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:47:33 crc kubenswrapper[4775]: I0321 04:47:33.531203 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:47:33 crc kubenswrapper[4775]: I0321 04:47:33.730621 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:33 crc kubenswrapper[4775]: I0321 04:47:33.731071 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:33 crc kubenswrapper[4775]: I0321 04:47:33.731340 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56"} Mar 21 04:47:33 crc kubenswrapper[4775]: I0321 04:47:33.731664 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:33 crc kubenswrapper[4775]: I0321 04:47:33.731698 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:33 crc kubenswrapper[4775]: I0321 04:47:33.731708 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:33 crc kubenswrapper[4775]: I0321 04:47:33.731754 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:33 crc kubenswrapper[4775]: I0321 04:47:33.731795 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:33 crc kubenswrapper[4775]: I0321 04:47:33.731806 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:33 crc kubenswrapper[4775]: I0321 04:47:33.744002 4775 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 21 04:47:34 crc kubenswrapper[4775]: I0321 04:47:34.046211 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:34 crc kubenswrapper[4775]: I0321 04:47:34.047568 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:34 crc kubenswrapper[4775]: I0321 04:47:34.047645 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:34 crc kubenswrapper[4775]: I0321 04:47:34.047659 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:34 crc kubenswrapper[4775]: I0321 04:47:34.047774 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 04:47:34 crc kubenswrapper[4775]: I0321 04:47:34.595711 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 04:47:34 crc kubenswrapper[4775]: I0321 04:47:34.595896 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:34 crc kubenswrapper[4775]: I0321 04:47:34.597500 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:34 crc kubenswrapper[4775]: I0321 04:47:34.597753 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:34 crc kubenswrapper[4775]: I0321 04:47:34.597944 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:34 crc kubenswrapper[4775]: I0321 04:47:34.732672 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:34 crc kubenswrapper[4775]: I0321 04:47:34.732698 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:34 crc kubenswrapper[4775]: I0321 04:47:34.733833 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:34 crc kubenswrapper[4775]: I0321 04:47:34.733860 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:34 crc kubenswrapper[4775]: I0321 04:47:34.733869 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:34 crc kubenswrapper[4775]: I0321 04:47:34.733864 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:34 crc kubenswrapper[4775]: I0321 04:47:34.734030 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:34 crc kubenswrapper[4775]: I0321 04:47:34.734056 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:35 crc kubenswrapper[4775]: I0321 04:47:35.016885 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:47:35 crc kubenswrapper[4775]: I0321 04:47:35.017414 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:35 crc kubenswrapper[4775]: I0321 04:47:35.018374 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:35 crc kubenswrapper[4775]: I0321 04:47:35.018407 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:35 crc kubenswrapper[4775]: I0321 04:47:35.018417 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:35 crc kubenswrapper[4775]: I0321 04:47:35.231861 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 21 04:47:35 crc kubenswrapper[4775]: I0321 04:47:35.394633 4775 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 21 04:47:35 crc kubenswrapper[4775]: I0321 04:47:35.394718 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 21 04:47:35 crc kubenswrapper[4775]: I0321 04:47:35.615299 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:47:35 crc kubenswrapper[4775]: I0321 04:47:35.621555 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:47:35 crc kubenswrapper[4775]: I0321 04:47:35.734868 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:35 crc kubenswrapper[4775]: I0321 04:47:35.734881 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:35 crc kubenswrapper[4775]: I0321 04:47:35.735870 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:35 crc kubenswrapper[4775]: I0321 04:47:35.735898 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:35 crc kubenswrapper[4775]: I0321 04:47:35.735908 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:35 crc kubenswrapper[4775]: I0321 04:47:35.735954 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:35 crc kubenswrapper[4775]: I0321 04:47:35.735987 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:35 crc kubenswrapper[4775]: I0321 04:47:35.735999 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:36 crc kubenswrapper[4775]: I0321 04:47:36.737466 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:36 crc kubenswrapper[4775]: I0321 04:47:36.738406 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:36 crc kubenswrapper[4775]: I0321 04:47:36.738468 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:36 crc kubenswrapper[4775]: I0321 04:47:36.738485 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:36 crc kubenswrapper[4775]: I0321 04:47:36.952706 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:47:37 crc kubenswrapper[4775]: E0321 04:47:37.732804 4775 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 04:47:37 crc kubenswrapper[4775]: I0321 04:47:37.738972 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:37 crc kubenswrapper[4775]: I0321 04:47:37.739733 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:37 crc kubenswrapper[4775]: I0321 04:47:37.739763 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:37 crc kubenswrapper[4775]: I0321 04:47:37.739775 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:40 crc kubenswrapper[4775]: I0321 04:47:40.646862 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 21 04:47:40 crc kubenswrapper[4775]: I0321 04:47:40.648540 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:40 crc kubenswrapper[4775]: I0321 04:47:40.650335 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:40 crc kubenswrapper[4775]: I0321 04:47:40.650439 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:40 crc kubenswrapper[4775]: I0321 04:47:40.650455 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:41 crc kubenswrapper[4775]: I0321 04:47:41.601178 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 21 04:47:41 crc kubenswrapper[4775]: E0321 04:47:41.897460 4775 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:47:41Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 21 04:47:41 crc kubenswrapper[4775]: W0321 04:47:41.899954 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:47:41Z is after 2026-02-23T05:33:13Z Mar 21 04:47:41 crc kubenswrapper[4775]: E0321 04:47:41.900015 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:47:41Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 21 04:47:41 crc kubenswrapper[4775]: W0321 04:47:41.900317 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:47:41Z is after 2026-02-23T05:33:13Z Mar 21 04:47:41 crc kubenswrapper[4775]: E0321 04:47:41.900451 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:47:41Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 21 04:47:41 crc kubenswrapper[4775]: E0321 04:47:41.901592 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:47:41Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 21 04:47:41 crc kubenswrapper[4775]: E0321 04:47:41.904148 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:47:41Z is after 2026-02-23T05:33:13Z" node="crc" Mar 21 04:47:41 crc kubenswrapper[4775]: W0321 04:47:41.905665 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:47:41Z is after 2026-02-23T05:33:13Z Mar 21 04:47:41 crc kubenswrapper[4775]: E0321 04:47:41.905747 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:47:41Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 21 04:47:41 crc kubenswrapper[4775]: I0321 04:47:41.907206 4775 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 21 04:47:41 crc kubenswrapper[4775]: I0321 04:47:41.907276 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 21 04:47:41 crc kubenswrapper[4775]: E0321 04:47:41.908397 4775 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:47:41Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189ec1da8a498a5d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.597660765 +0000 UTC m=+0.574124389,LastTimestamp:2026-03-21 04:47:27.597660765 +0000 UTC m=+0.574124389,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:41 crc kubenswrapper[4775]: W0321 04:47:41.909674 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:47:41Z is after 2026-02-23T05:33:13Z Mar 21 04:47:41 crc kubenswrapper[4775]: E0321 04:47:41.909747 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:47:41Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 21 04:47:41 crc kubenswrapper[4775]: I0321 04:47:41.913815 4775 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 21 04:47:41 crc kubenswrapper[4775]: I0321 04:47:41.913883 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 21 04:47:42 crc kubenswrapper[4775]: I0321 04:47:42.607019 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:47:42Z is after 2026-02-23T05:33:13Z Mar 21 04:47:42 crc kubenswrapper[4775]: I0321 04:47:42.755109 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 21 04:47:42 crc kubenswrapper[4775]: I0321 04:47:42.757734 4775 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c38b09928e8a15368f55622732324802c7af7bc3d13c7cf2d1d00b4ac43fef94" exitCode=255 Mar 21 04:47:42 crc kubenswrapper[4775]: I0321 04:47:42.757790 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"c38b09928e8a15368f55622732324802c7af7bc3d13c7cf2d1d00b4ac43fef94"} Mar 21 04:47:42 crc kubenswrapper[4775]: I0321 04:47:42.757959 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:42 crc kubenswrapper[4775]: I0321 04:47:42.758944 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:42 crc kubenswrapper[4775]: I0321 04:47:42.759049 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:42 crc kubenswrapper[4775]: I0321 04:47:42.759139 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:42 crc kubenswrapper[4775]: I0321 04:47:42.759735 4775 scope.go:117] "RemoveContainer" containerID="c38b09928e8a15368f55622732324802c7af7bc3d13c7cf2d1d00b4ac43fef94" Mar 21 04:47:43 crc kubenswrapper[4775]: I0321 04:47:43.535717 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:47:43 crc kubenswrapper[4775]: I0321 04:47:43.602681 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:47:43Z is after 2026-02-23T05:33:13Z Mar 21 04:47:43 crc kubenswrapper[4775]: I0321 04:47:43.763683 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 21 04:47:43 crc kubenswrapper[4775]: I0321 04:47:43.766339 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a3ec38c653682ae3ceeecec7bf83cf11591c8de9989f8ce7f4b2ce4297b763cc"} Mar 21 04:47:43 crc kubenswrapper[4775]: I0321 04:47:43.766560 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:43 crc kubenswrapper[4775]: I0321 04:47:43.768084 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:43 crc kubenswrapper[4775]: I0321 04:47:43.768143 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:43 crc kubenswrapper[4775]: I0321 04:47:43.768155 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:43 crc kubenswrapper[4775]: I0321 04:47:43.771347 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:47:44 crc kubenswrapper[4775]: I0321 04:47:44.602748 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:47:44Z is after 2026-02-23T05:33:13Z Mar 21 04:47:44 crc kubenswrapper[4775]: I0321 04:47:44.770481 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 21 04:47:44 crc kubenswrapper[4775]: I0321 04:47:44.771214 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 21 04:47:44 crc kubenswrapper[4775]: I0321 04:47:44.772761 4775 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a3ec38c653682ae3ceeecec7bf83cf11591c8de9989f8ce7f4b2ce4297b763cc" exitCode=255 Mar 21 04:47:44 crc kubenswrapper[4775]: I0321 04:47:44.772901 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a3ec38c653682ae3ceeecec7bf83cf11591c8de9989f8ce7f4b2ce4297b763cc"} Mar 21 04:47:44 crc kubenswrapper[4775]: I0321 04:47:44.773013 4775 scope.go:117] "RemoveContainer" containerID="c38b09928e8a15368f55622732324802c7af7bc3d13c7cf2d1d00b4ac43fef94" Mar 21 04:47:44 crc kubenswrapper[4775]: I0321 04:47:44.773245 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:44 crc kubenswrapper[4775]: I0321 04:47:44.774259 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:44 crc kubenswrapper[4775]: I0321 04:47:44.774368 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:44 crc kubenswrapper[4775]: I0321 04:47:44.774438 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:44 crc kubenswrapper[4775]: I0321 04:47:44.774965 4775 scope.go:117] "RemoveContainer" containerID="a3ec38c653682ae3ceeecec7bf83cf11591c8de9989f8ce7f4b2ce4297b763cc" Mar 21 04:47:44 crc kubenswrapper[4775]: E0321 04:47:44.775224 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 04:47:45 crc kubenswrapper[4775]: I0321 04:47:45.022039 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:47:45 crc kubenswrapper[4775]: I0321 04:47:45.022198 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:45 crc kubenswrapper[4775]: I0321 04:47:45.023131 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:45 crc kubenswrapper[4775]: I0321 04:47:45.023160 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:45 crc kubenswrapper[4775]: I0321 04:47:45.023170 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:45 crc kubenswrapper[4775]: I0321 04:47:45.395720 4775 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 21 04:47:45 crc kubenswrapper[4775]: I0321 04:47:45.395843 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 21 04:47:45 crc kubenswrapper[4775]: I0321 04:47:45.604565 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:47:45Z is after 2026-02-23T05:33:13Z Mar 21 04:47:45 crc kubenswrapper[4775]: I0321 04:47:45.778386 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 21 04:47:45 crc kubenswrapper[4775]: I0321 04:47:45.780553 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:45 crc kubenswrapper[4775]: I0321 04:47:45.781436 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:45 crc kubenswrapper[4775]: I0321 04:47:45.781478 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:45 crc kubenswrapper[4775]: I0321 04:47:45.781492 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:45 crc kubenswrapper[4775]: I0321 04:47:45.782135 4775 scope.go:117] "RemoveContainer" containerID="a3ec38c653682ae3ceeecec7bf83cf11591c8de9989f8ce7f4b2ce4297b763cc" Mar 21 04:47:45 crc kubenswrapper[4775]: E0321 04:47:45.782378 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 04:47:46 crc kubenswrapper[4775]: I0321 04:47:46.602980 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:47:46Z is after 2026-02-23T05:33:13Z Mar 21 04:47:47 crc kubenswrapper[4775]: I0321 04:47:47.603665 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:47:47Z is after 2026-02-23T05:33:13Z Mar 21 04:47:47 crc kubenswrapper[4775]: E0321 04:47:47.732947 4775 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 04:47:48 crc kubenswrapper[4775]: I0321 04:47:48.304909 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:48 crc kubenswrapper[4775]: I0321 04:47:48.306047 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:48 crc kubenswrapper[4775]: I0321 04:47:48.306107 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:48 crc kubenswrapper[4775]: I0321 04:47:48.306143 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:48 crc kubenswrapper[4775]: I0321 04:47:48.306171 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 04:47:48 crc kubenswrapper[4775]: E0321 04:47:48.306594 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 04:47:48 crc kubenswrapper[4775]: E0321 04:47:48.309583 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 04:47:48 crc kubenswrapper[4775]: I0321 04:47:48.604818 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:47:48 crc kubenswrapper[4775]: W0321 04:47:48.838603 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 21 04:47:48 crc kubenswrapper[4775]: E0321 04:47:48.838685 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 21 04:47:48 crc kubenswrapper[4775]: W0321 04:47:48.916321 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 21 04:47:48 crc kubenswrapper[4775]: E0321 04:47:48.916386 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 21 04:47:49 crc kubenswrapper[4775]: I0321 04:47:49.605353 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:47:49 crc kubenswrapper[4775]: I0321 04:47:49.703335 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:47:49 crc kubenswrapper[4775]: I0321 04:47:49.703524 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:49 crc kubenswrapper[4775]: I0321 04:47:49.704621 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:49 crc kubenswrapper[4775]: I0321 04:47:49.704656 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:49 crc kubenswrapper[4775]: I0321 04:47:49.704667 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:49 crc kubenswrapper[4775]: I0321 04:47:49.705223 4775 scope.go:117] "RemoveContainer" containerID="a3ec38c653682ae3ceeecec7bf83cf11591c8de9989f8ce7f4b2ce4297b763cc" Mar 21 04:47:49 crc kubenswrapper[4775]: E0321 04:47:49.705411 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 04:47:50 crc kubenswrapper[4775]: I0321 04:47:50.231415 4775 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 21 04:47:50 crc kubenswrapper[4775]: I0321 04:47:50.244752 4775 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 21 04:47:50 crc kubenswrapper[4775]: W0321 04:47:50.330804 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 21 04:47:50 crc kubenswrapper[4775]: E0321 04:47:50.330913 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 21 04:47:50 crc kubenswrapper[4775]: I0321 04:47:50.603180 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:47:50 crc kubenswrapper[4775]: I0321 04:47:50.685754 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 21 04:47:50 crc kubenswrapper[4775]: I0321 04:47:50.686074 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:50 crc kubenswrapper[4775]: I0321 04:47:50.687749 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:50 crc kubenswrapper[4775]: I0321 04:47:50.687802 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:50 crc kubenswrapper[4775]: I0321 04:47:50.687819 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:50 crc kubenswrapper[4775]: I0321 04:47:50.703753 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 21 04:47:50 crc kubenswrapper[4775]: I0321 04:47:50.793372 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:50 crc kubenswrapper[4775]: I0321 04:47:50.794767 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:50 crc kubenswrapper[4775]: I0321 04:47:50.794841 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:50 crc kubenswrapper[4775]: I0321 04:47:50.794865 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:51 crc kubenswrapper[4775]: I0321 04:47:51.603800 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:47:51 crc kubenswrapper[4775]: E0321 04:47:51.913746 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8a498a5d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.597660765 +0000 UTC m=+0.574124389,LastTimestamp:2026-03-21 04:47:27.597660765 +0000 UTC m=+0.574124389,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:51 crc kubenswrapper[4775]: E0321 04:47:51.919201 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d0b56 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653464918 +0000 UTC m=+0.629928542,LastTimestamp:2026-03-21 04:47:27.653464918 +0000 UTC m=+0.629928542,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:51 crc kubenswrapper[4775]: E0321 04:47:51.923644 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d6092 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653486738 +0000 UTC m=+0.629950362,LastTimestamp:2026-03-21 04:47:27.653486738 +0000 UTC m=+0.629950362,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:51 crc kubenswrapper[4775]: E0321 04:47:51.927720 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d84c9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653496009 +0000 UTC m=+0.629959633,LastTimestamp:2026-03-21 04:47:27.653496009 +0000 UTC m=+0.629959633,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:51 crc kubenswrapper[4775]: E0321 04:47:51.931823 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da9215a278 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.728476792 +0000 UTC m=+0.704940426,LastTimestamp:2026-03-21 04:47:27.728476792 +0000 UTC m=+0.704940426,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:51 crc kubenswrapper[4775]: E0321 04:47:51.937371 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ec1da8d9d0b56\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d0b56 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653464918 +0000 UTC m=+0.629928542,LastTimestamp:2026-03-21 04:47:27.762896765 +0000 UTC m=+0.739360399,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:51 crc kubenswrapper[4775]: E0321 04:47:51.941528 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ec1da8d9d6092\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d6092 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653486738 +0000 UTC m=+0.629950362,LastTimestamp:2026-03-21 04:47:27.762922126 +0000 UTC m=+0.739385760,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:51 crc kubenswrapper[4775]: E0321 04:47:51.945727 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ec1da8d9d84c9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d84c9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653496009 +0000 UTC m=+0.629959633,LastTimestamp:2026-03-21 04:47:27.762933036 +0000 UTC m=+0.739396670,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:51 crc kubenswrapper[4775]: E0321 04:47:51.950468 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ec1da8d9d0b56\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d0b56 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653464918 +0000 UTC m=+0.629928542,LastTimestamp:2026-03-21 04:47:27.763968134 +0000 UTC m=+0.740431768,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:51 crc kubenswrapper[4775]: E0321 04:47:51.954416 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ec1da8d9d6092\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d6092 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653486738 +0000 UTC m=+0.629950362,LastTimestamp:2026-03-21 04:47:27.763987255 +0000 UTC m=+0.740450889,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:51 crc kubenswrapper[4775]: E0321 04:47:51.960075 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ec1da8d9d84c9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d84c9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653496009 +0000 UTC m=+0.629959633,LastTimestamp:2026-03-21 04:47:27.763997745 +0000 UTC m=+0.740461379,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:51 crc kubenswrapper[4775]: E0321 04:47:51.967466 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ec1da8d9d0b56\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d0b56 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653464918 +0000 UTC m=+0.629928542,LastTimestamp:2026-03-21 04:47:27.766332088 +0000 UTC m=+0.742795722,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:51 crc kubenswrapper[4775]: E0321 04:47:51.972549 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ec1da8d9d6092\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d6092 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653486738 +0000 UTC m=+0.629950362,LastTimestamp:2026-03-21 04:47:27.766348809 +0000 UTC m=+0.742812443,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:51 crc kubenswrapper[4775]: E0321 04:47:51.977530 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ec1da8d9d84c9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d84c9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653496009 +0000 UTC m=+0.629959633,LastTimestamp:2026-03-21 04:47:27.766361109 +0000 UTC m=+0.742824743,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:51 crc kubenswrapper[4775]: E0321 04:47:51.981955 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ec1da8d9d0b56\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d0b56 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653464918 +0000 UTC m=+0.629928542,LastTimestamp:2026-03-21 04:47:27.767301695 +0000 UTC m=+0.743765319,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:51 crc kubenswrapper[4775]: E0321 04:47:51.985877 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ec1da8d9d6092\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d6092 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653486738 +0000 UTC m=+0.629950362,LastTimestamp:2026-03-21 04:47:27.767318645 +0000 UTC m=+0.743782269,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:51 crc kubenswrapper[4775]: E0321 04:47:51.990999 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ec1da8d9d84c9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d84c9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653496009 +0000 UTC m=+0.629959633,LastTimestamp:2026-03-21 04:47:27.767331535 +0000 UTC m=+0.743795159,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:51 crc kubenswrapper[4775]: E0321 04:47:51.995851 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ec1da8d9d0b56\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d0b56 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653464918 +0000 UTC m=+0.629928542,LastTimestamp:2026-03-21 04:47:27.76748279 +0000 UTC m=+0.743946424,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.000162 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ec1da8d9d6092\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d6092 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653486738 +0000 UTC m=+0.629950362,LastTimestamp:2026-03-21 04:47:27.76749861 +0000 UTC m=+0.743962254,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.003824 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ec1da8d9d84c9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d84c9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653496009 +0000 UTC m=+0.629959633,LastTimestamp:2026-03-21 04:47:27.76751009 +0000 UTC m=+0.743973724,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.008318 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ec1da8d9d0b56\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d0b56 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653464918 +0000 UTC m=+0.629928542,LastTimestamp:2026-03-21 04:47:27.767684555 +0000 UTC m=+0.744148179,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.012217 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ec1da8d9d6092\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d6092 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653486738 +0000 UTC m=+0.629950362,LastTimestamp:2026-03-21 04:47:27.767735226 +0000 UTC m=+0.744198850,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.015648 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ec1da8d9d84c9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d84c9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653496009 +0000 UTC m=+0.629959633,LastTimestamp:2026-03-21 04:47:27.767761947 +0000 UTC m=+0.744225571,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.020241 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ec1da8d9d0b56\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d0b56 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653464918 +0000 UTC m=+0.629928542,LastTimestamp:2026-03-21 04:47:27.768414995 +0000 UTC m=+0.744878629,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.025396 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ec1da8d9d6092\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ec1da8d9d6092 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:27.653486738 +0000 UTC m=+0.629950362,LastTimestamp:2026-03-21 04:47:27.768428985 +0000 UTC m=+0.744892619,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.031185 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189ec1daacc2bcfe openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:28.176028926 +0000 UTC m=+1.152492550,LastTimestamp:2026-03-21 04:47:28.176028926 +0000 UTC m=+1.152492550,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.035830 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ec1daacce7940 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:28.176798016 +0000 UTC m=+1.153261640,LastTimestamp:2026-03-21 04:47:28.176798016 +0000 UTC m=+1.153261640,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.041214 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1daacd022c5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:28.176906949 +0000 UTC m=+1.153370563,LastTimestamp:2026-03-21 04:47:28.176906949 +0000 UTC m=+1.153370563,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.046513 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1daad5d6e4b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:28.186166859 +0000 UTC m=+1.162630483,LastTimestamp:2026-03-21 04:47:28.186166859 +0000 UTC m=+1.162630483,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.050496 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ec1daae2e97d6 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:28.199874518 +0000 UTC m=+1.176338142,LastTimestamp:2026-03-21 04:47:28.199874518 +0000 UTC m=+1.176338142,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.054752 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1dacee57213 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:28.748728851 +0000 UTC m=+1.725192475,LastTimestamp:2026-03-21 04:47:28.748728851 +0000 UTC m=+1.725192475,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.058463 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1dacee6d6ff openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:28.748820223 +0000 UTC m=+1.725283857,LastTimestamp:2026-03-21 04:47:28.748820223 +0000 UTC m=+1.725283857,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.061995 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189ec1dacee7af87 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:28.748875655 +0000 UTC m=+1.725339279,LastTimestamp:2026-03-21 04:47:28.748875655 +0000 UTC m=+1.725339279,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.065286 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ec1dacee9cd06 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:28.749014278 +0000 UTC m=+1.725477902,LastTimestamp:2026-03-21 04:47:28.749014278 +0000 UTC m=+1.725477902,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.069786 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ec1daceeac6ea openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:28.74907825 +0000 UTC m=+1.725541874,LastTimestamp:2026-03-21 04:47:28.74907825 +0000 UTC m=+1.725541874,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.073928 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ec1dacf77de53 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:28.758324819 +0000 UTC m=+1.734788443,LastTimestamp:2026-03-21 04:47:28.758324819 +0000 UTC m=+1.734788443,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.078320 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ec1dacf967b9b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:28.760331163 +0000 UTC m=+1.736794787,LastTimestamp:2026-03-21 04:47:28.760331163 +0000 UTC m=+1.736794787,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.081732 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189ec1dacf970da0 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:28.760368544 +0000 UTC m=+1.736832168,LastTimestamp:2026-03-21 04:47:28.760368544 +0000 UTC m=+1.736832168,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.085430 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ec1dacfa25430 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:28.761107504 +0000 UTC m=+1.737571118,LastTimestamp:2026-03-21 04:47:28.761107504 +0000 UTC m=+1.737571118,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.089058 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1dacfa91ce0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:28.761552096 +0000 UTC m=+1.738015720,LastTimestamp:2026-03-21 04:47:28.761552096 +0000 UTC m=+1.738015720,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.092907 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1dacfa9571f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:28.761567007 +0000 UTC m=+1.738030631,LastTimestamp:2026-03-21 04:47:28.761567007 +0000 UTC m=+1.738030631,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.097656 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ec1dae0e03574 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.05037554 +0000 UTC m=+2.026839174,LastTimestamp:2026-03-21 04:47:29.05037554 +0000 UTC m=+2.026839174,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.101987 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ec1dae17e77a2 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.06074717 +0000 UTC m=+2.037210794,LastTimestamp:2026-03-21 04:47:29.06074717 +0000 UTC m=+2.037210794,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.106051 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ec1dae18aadf7 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.061547511 +0000 UTC m=+2.038011135,LastTimestamp:2026-03-21 04:47:29.061547511 +0000 UTC m=+2.038011135,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.109590 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ec1daeb936f02 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.229893378 +0000 UTC m=+2.206357002,LastTimestamp:2026-03-21 04:47:29.229893378 +0000 UTC m=+2.206357002,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.113079 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ec1daecc8d0ad openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.250169005 +0000 UTC m=+2.226632629,LastTimestamp:2026-03-21 04:47:29.250169005 +0000 UTC m=+2.226632629,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.116724 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ec1daecd5c6ee openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.251018478 +0000 UTC m=+2.227482102,LastTimestamp:2026-03-21 04:47:29.251018478 +0000 UTC m=+2.227482102,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.120040 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ec1daf5c7fe7e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.401110142 +0000 UTC m=+2.377573766,LastTimestamp:2026-03-21 04:47:29.401110142 +0000 UTC m=+2.377573766,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.123343 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ec1daf6ee5492 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.420399762 +0000 UTC m=+2.396863406,LastTimestamp:2026-03-21 04:47:29.420399762 +0000 UTC m=+2.396863406,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.126754 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1db061f1364 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.67525258 +0000 UTC m=+2.651716204,LastTimestamp:2026-03-21 04:47:29.67525258 +0000 UTC m=+2.651716204,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.131068 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1db063fcf08 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.677397768 +0000 UTC m=+2.653861392,LastTimestamp:2026-03-21 04:47:29.677397768 +0000 UTC m=+2.653861392,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.134538 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ec1db067f1426 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.68154423 +0000 UTC m=+2.658007864,LastTimestamp:2026-03-21 04:47:29.68154423 +0000 UTC m=+2.658007864,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.138813 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189ec1db06a2ac81 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.683876993 +0000 UTC m=+2.660340617,LastTimestamp:2026-03-21 04:47:29.683876993 +0000 UTC m=+2.660340617,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.143360 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1db1343f91e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.895774494 +0000 UTC m=+2.872238118,LastTimestamp:2026-03-21 04:47:29.895774494 +0000 UTC m=+2.872238118,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.147283 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1db13582ab5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.897097909 +0000 UTC m=+2.873561553,LastTimestamp:2026-03-21 04:47:29.897097909 +0000 UTC m=+2.873561553,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.152642 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ec1db135ff505 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.897608453 +0000 UTC m=+2.874072077,LastTimestamp:2026-03-21 04:47:29.897608453 +0000 UTC m=+2.874072077,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.156849 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189ec1db13612f49 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.897688905 +0000 UTC m=+2.874152529,LastTimestamp:2026-03-21 04:47:29.897688905 +0000 UTC m=+2.874152529,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.162143 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189ec1db14ad2120 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.919443232 +0000 UTC m=+2.895906856,LastTimestamp:2026-03-21 04:47:29.919443232 +0000 UTC m=+2.895906856,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.166053 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ec1db14b0aa12 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.919674898 +0000 UTC m=+2.896138522,LastTimestamp:2026-03-21 04:47:29.919674898 +0000 UTC m=+2.896138522,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.169539 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ec1db14bf85ba openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.920648634 +0000 UTC m=+2.897112258,LastTimestamp:2026-03-21 04:47:29.920648634 +0000 UTC m=+2.897112258,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.172816 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1db157d2ad9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.933077209 +0000 UTC m=+2.909540833,LastTimestamp:2026-03-21 04:47:29.933077209 +0000 UTC m=+2.909540833,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.176072 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1db1594a611 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.934616081 +0000 UTC m=+2.911079705,LastTimestamp:2026-03-21 04:47:29.934616081 +0000 UTC m=+2.911079705,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.180325 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1db161c65b4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.9435125 +0000 UTC m=+2.919976124,LastTimestamp:2026-03-21 04:47:29.9435125 +0000 UTC m=+2.919976124,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.185546 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1db2725c9bc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.229340604 +0000 UTC m=+3.205804228,LastTimestamp:2026-03-21 04:47:30.229340604 +0000 UTC m=+3.205804228,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.189076 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ec1db27874301 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.235728641 +0000 UTC m=+3.212192265,LastTimestamp:2026-03-21 04:47:30.235728641 +0000 UTC m=+3.212192265,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.192648 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1db27da0019 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.241151001 +0000 UTC m=+3.217614625,LastTimestamp:2026-03-21 04:47:30.241151001 +0000 UTC m=+3.217614625,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.196692 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1db27e4bf4e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.24185531 +0000 UTC m=+3.218318934,LastTimestamp:2026-03-21 04:47:30.24185531 +0000 UTC m=+3.218318934,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.201011 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ec1db28641015 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.250199061 +0000 UTC m=+3.226662685,LastTimestamp:2026-03-21 04:47:30.250199061 +0000 UTC m=+3.226662685,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.205229 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ec1db28740eb0 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.25124728 +0000 UTC m=+3.227710984,LastTimestamp:2026-03-21 04:47:30.25124728 +0000 UTC m=+3.227710984,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.208686 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1db31ce6f4f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.408165199 +0000 UTC m=+3.384628823,LastTimestamp:2026-03-21 04:47:30.408165199 +0000 UTC m=+3.384628823,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.213298 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ec1db32028d01 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.411580673 +0000 UTC m=+3.388044307,LastTimestamp:2026-03-21 04:47:30.411580673 +0000 UTC m=+3.388044307,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.216839 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1db336753ea openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.43496241 +0000 UTC m=+3.411426034,LastTimestamp:2026-03-21 04:47:30.43496241 +0000 UTC m=+3.411426034,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.221211 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1db33783e77 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.436071031 +0000 UTC m=+3.412534655,LastTimestamp:2026-03-21 04:47:30.436071031 +0000 UTC m=+3.412534655,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.225008 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ec1db33dfb810 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.442852368 +0000 UTC m=+3.419315992,LastTimestamp:2026-03-21 04:47:30.442852368 +0000 UTC m=+3.419315992,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.229482 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1db3d0ae7a5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.596677541 +0000 UTC m=+3.573141165,LastTimestamp:2026-03-21 04:47:30.596677541 +0000 UTC m=+3.573141165,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.233536 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1db3dd9c610 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.610234896 +0000 UTC m=+3.586698530,LastTimestamp:2026-03-21 04:47:30.610234896 +0000 UTC m=+3.586698530,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.237801 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1db3deb28d0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.611374288 +0000 UTC m=+3.587837922,LastTimestamp:2026-03-21 04:47:30.611374288 +0000 UTC m=+3.587837922,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.241884 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1db42d02fd1 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.693492689 +0000 UTC m=+3.669956313,LastTimestamp:2026-03-21 04:47:30.693492689 +0000 UTC m=+3.669956313,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.246407 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1db46e308fe openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.761836798 +0000 UTC m=+3.738300422,LastTimestamp:2026-03-21 04:47:30.761836798 +0000 UTC m=+3.738300422,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.249872 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1db47f0ca4b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.779515467 +0000 UTC m=+3.755979091,LastTimestamp:2026-03-21 04:47:30.779515467 +0000 UTC m=+3.755979091,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.253939 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1db4d4456c6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.868876998 +0000 UTC m=+3.845340622,LastTimestamp:2026-03-21 04:47:30.868876998 +0000 UTC m=+3.845340622,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.257025 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1db4e2f08c9 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.884257993 +0000 UTC m=+3.860721617,LastTimestamp:2026-03-21 04:47:30.884257993 +0000 UTC m=+3.860721617,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.261343 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1db802c518e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:31.722940814 +0000 UTC m=+4.699404478,LastTimestamp:2026-03-21 04:47:31.722940814 +0000 UTC m=+4.699404478,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.264808 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1db8d05e23b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:31.938525755 +0000 UTC m=+4.914989379,LastTimestamp:2026-03-21 04:47:31.938525755 +0000 UTC m=+4.914989379,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.268244 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1db8db7d854 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:31.950188628 +0000 UTC m=+4.926652262,LastTimestamp:2026-03-21 04:47:31.950188628 +0000 UTC m=+4.926652262,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.271728 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1db8dc59971 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:31.951090033 +0000 UTC m=+4.927553657,LastTimestamp:2026-03-21 04:47:31.951090033 +0000 UTC m=+4.927553657,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.275078 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1db9b63a4bc openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:32.17955142 +0000 UTC m=+5.156015074,LastTimestamp:2026-03-21 04:47:32.17955142 +0000 UTC m=+5.156015074,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.278661 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1db9c4f3799 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:32.194989977 +0000 UTC m=+5.171453641,LastTimestamp:2026-03-21 04:47:32.194989977 +0000 UTC m=+5.171453641,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.282135 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1db9c76cd0b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:32.197584139 +0000 UTC m=+5.174047763,LastTimestamp:2026-03-21 04:47:32.197584139 +0000 UTC m=+5.174047763,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.285663 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1dba7d409b3 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:32.388243891 +0000 UTC m=+5.364707515,LastTimestamp:2026-03-21 04:47:32.388243891 +0000 UTC m=+5.364707515,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.289336 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1dba8dd36e5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:32.405622501 +0000 UTC m=+5.382086135,LastTimestamp:2026-03-21 04:47:32.405622501 +0000 UTC m=+5.382086135,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.292285 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1dba8eebe21 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:32.406771233 +0000 UTC m=+5.383234857,LastTimestamp:2026-03-21 04:47:32.406771233 +0000 UTC m=+5.383234857,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.295877 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1dbb335a68b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:32.579190411 +0000 UTC m=+5.555654035,LastTimestamp:2026-03-21 04:47:32.579190411 +0000 UTC m=+5.555654035,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.299535 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1dbb3e4e968 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:32.590676328 +0000 UTC m=+5.567139942,LastTimestamp:2026-03-21 04:47:32.590676328 +0000 UTC m=+5.567139942,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.303406 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1dbb3f41170 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:32.591669616 +0000 UTC m=+5.568133240,LastTimestamp:2026-03-21 04:47:32.591669616 +0000 UTC m=+5.568133240,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.307403 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1dbbccfeddb openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:32.740296155 +0000 UTC m=+5.716759799,LastTimestamp:2026-03-21 04:47:32.740296155 +0000 UTC m=+5.716759799,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.311288 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ec1dbbda7e3a5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:32.754449317 +0000 UTC m=+5.730912941,LastTimestamp:2026-03-21 04:47:32.754449317 +0000 UTC m=+5.730912941,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.315352 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 21 04:47:52 crc kubenswrapper[4775]: &Event{ObjectMeta:{kube-controller-manager-crc.189ec1dc5b06d854 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 21 04:47:52 crc kubenswrapper[4775]: body: Mar 21 04:47:52 crc kubenswrapper[4775]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:35.394695252 +0000 UTC m=+8.371158876,LastTimestamp:2026-03-21 04:47:35.394695252 +0000 UTC m=+8.371158876,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 21 04:47:52 crc kubenswrapper[4775]: > Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.318705 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ec1dc5b07b0b4 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:35.394750644 +0000 UTC m=+8.371214268,LastTimestamp:2026-03-21 04:47:35.394750644 +0000 UTC m=+8.371214268,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.323617 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 21 04:47:52 crc kubenswrapper[4775]: &Event{ObjectMeta:{kube-apiserver-crc.189ec1dddf349906 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 21 04:47:52 crc kubenswrapper[4775]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 21 04:47:52 crc kubenswrapper[4775]: Mar 21 04:47:52 crc kubenswrapper[4775]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:41.90725351 +0000 UTC m=+14.883717124,LastTimestamp:2026-03-21 04:47:41.90725351 +0000 UTC m=+14.883717124,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 21 04:47:52 crc kubenswrapper[4775]: > Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.329193 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1dddf360eb5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:41.907349173 +0000 UTC m=+14.883812817,LastTimestamp:2026-03-21 04:47:41.907349173 +0000 UTC m=+14.883812817,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.333415 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189ec1dddf349906\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 21 04:47:52 crc kubenswrapper[4775]: &Event{ObjectMeta:{kube-apiserver-crc.189ec1dddf349906 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 21 04:47:52 crc kubenswrapper[4775]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 21 04:47:52 crc kubenswrapper[4775]: Mar 21 04:47:52 crc kubenswrapper[4775]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:41.90725351 +0000 UTC m=+14.883717124,LastTimestamp:2026-03-21 04:47:41.913863893 +0000 UTC m=+14.890327517,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 21 04:47:52 crc kubenswrapper[4775]: > Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.336409 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189ec1dddf360eb5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1dddf360eb5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:41.907349173 +0000 UTC m=+14.883812817,LastTimestamp:2026-03-21 04:47:41.913906864 +0000 UTC m=+14.890370488,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.339650 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189ec1db3deb28d0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1db3deb28d0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.611374288 +0000 UTC m=+3.587837922,LastTimestamp:2026-03-21 04:47:42.760975897 +0000 UTC m=+15.737439521,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.344178 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189ec1db46e308fe\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1db46e308fe openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.761836798 +0000 UTC m=+3.738300422,LastTimestamp:2026-03-21 04:47:43.003957656 +0000 UTC m=+15.980421280,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.349706 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189ec1db47f0ca4b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ec1db47f0ca4b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:30.779515467 +0000 UTC m=+3.755979091,LastTimestamp:2026-03-21 04:47:43.012821501 +0000 UTC m=+15.989285125,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.355530 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189ec1dc5b06d854\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 21 04:47:52 crc kubenswrapper[4775]: &Event{ObjectMeta:{kube-controller-manager-crc.189ec1dc5b06d854 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 21 04:47:52 crc kubenswrapper[4775]: body: Mar 21 04:47:52 crc kubenswrapper[4775]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:35.394695252 +0000 UTC m=+8.371158876,LastTimestamp:2026-03-21 04:47:45.395807312 +0000 UTC m=+18.372270936,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 21 04:47:52 crc kubenswrapper[4775]: > Mar 21 04:47:52 crc kubenswrapper[4775]: E0321 04:47:52.360306 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189ec1dc5b07b0b4\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ec1dc5b07b0b4 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:35.394750644 +0000 UTC m=+8.371214268,LastTimestamp:2026-03-21 04:47:45.395884935 +0000 UTC m=+18.372348559,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:52 crc kubenswrapper[4775]: I0321 04:47:52.604354 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:47:53 crc kubenswrapper[4775]: I0321 04:47:53.132924 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:47:53 crc kubenswrapper[4775]: I0321 04:47:53.133191 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:53 crc kubenswrapper[4775]: I0321 04:47:53.134492 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:53 crc kubenswrapper[4775]: I0321 04:47:53.134556 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:53 crc kubenswrapper[4775]: I0321 04:47:53.134569 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:53 crc kubenswrapper[4775]: I0321 04:47:53.135108 4775 scope.go:117] "RemoveContainer" containerID="a3ec38c653682ae3ceeecec7bf83cf11591c8de9989f8ce7f4b2ce4297b763cc" Mar 21 04:47:53 crc kubenswrapper[4775]: E0321 04:47:53.135286 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 04:47:53 crc kubenswrapper[4775]: W0321 04:47:53.150994 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 21 04:47:53 crc kubenswrapper[4775]: E0321 04:47:53.151077 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 21 04:47:53 crc kubenswrapper[4775]: I0321 04:47:53.604900 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:47:54 crc kubenswrapper[4775]: I0321 04:47:54.603991 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.309748 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.310926 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.310981 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.311012 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.311047 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 04:47:55 crc kubenswrapper[4775]: E0321 04:47:55.311443 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 04:47:55 crc kubenswrapper[4775]: E0321 04:47:55.315278 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.394830 4775 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.394931 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.395090 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.395265 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.396258 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.396296 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.396305 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.396733 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"851dab44c58218d203e180465f8a216bfeec8271e78102856d11545520feabb4"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.396874 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://851dab44c58218d203e180465f8a216bfeec8271e78102856d11545520feabb4" gracePeriod=30 Mar 21 04:47:55 crc kubenswrapper[4775]: E0321 04:47:55.399958 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189ec1dc5b06d854\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 21 04:47:55 crc kubenswrapper[4775]: &Event{ObjectMeta:{kube-controller-manager-crc.189ec1dc5b06d854 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 21 04:47:55 crc kubenswrapper[4775]: body: Mar 21 04:47:55 crc kubenswrapper[4775]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:35.394695252 +0000 UTC m=+8.371158876,LastTimestamp:2026-03-21 04:47:55.394900916 +0000 UTC m=+28.371364540,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 21 04:47:55 crc kubenswrapper[4775]: > Mar 21 04:47:55 crc kubenswrapper[4775]: E0321 04:47:55.405483 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189ec1dc5b07b0b4\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ec1dc5b07b0b4 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:35.394750644 +0000 UTC m=+8.371214268,LastTimestamp:2026-03-21 04:47:55.39503712 +0000 UTC m=+28.371500744,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:55 crc kubenswrapper[4775]: E0321 04:47:55.411406 4775 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ec1e1033fad92 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:55.39686133 +0000 UTC m=+28.373324954,LastTimestamp:2026-03-21 04:47:55.39686133 +0000 UTC m=+28.373324954,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:55 crc kubenswrapper[4775]: E0321 04:47:55.514533 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189ec1dacfa25430\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ec1dacfa25430 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:28.761107504 +0000 UTC m=+1.737571118,LastTimestamp:2026-03-21 04:47:55.509932247 +0000 UTC m=+28.486395871,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.603339 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:47:55 crc kubenswrapper[4775]: E0321 04:47:55.649655 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189ec1dae0e03574\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ec1dae0e03574 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.05037554 +0000 UTC m=+2.026839174,LastTimestamp:2026-03-21 04:47:55.64386567 +0000 UTC m=+28.620329294,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:55 crc kubenswrapper[4775]: E0321 04:47:55.657308 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189ec1dae17e77a2\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ec1dae17e77a2 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:29.06074717 +0000 UTC m=+2.037210794,LastTimestamp:2026-03-21 04:47:55.65324367 +0000 UTC m=+28.629707294,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.806405 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.806978 4775 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="851dab44c58218d203e180465f8a216bfeec8271e78102856d11545520feabb4" exitCode=255 Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.807092 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"851dab44c58218d203e180465f8a216bfeec8271e78102856d11545520feabb4"} Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.807227 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e5615480a5b452c6bce19fa04738bb5d68984f62258c9486fc2ae787cfb9b7fd"} Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.807612 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.808880 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.809042 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:55 crc kubenswrapper[4775]: I0321 04:47:55.809165 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:56 crc kubenswrapper[4775]: I0321 04:47:56.604302 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:47:56 crc kubenswrapper[4775]: I0321 04:47:56.953707 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:47:56 crc kubenswrapper[4775]: I0321 04:47:56.953862 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:47:56 crc kubenswrapper[4775]: I0321 04:47:56.955044 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:47:56 crc kubenswrapper[4775]: I0321 04:47:56.955094 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:47:56 crc kubenswrapper[4775]: I0321 04:47:56.955111 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:47:57 crc kubenswrapper[4775]: I0321 04:47:57.603909 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:47:57 crc kubenswrapper[4775]: E0321 04:47:57.733148 4775 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 04:47:58 crc kubenswrapper[4775]: I0321 04:47:58.604429 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:47:59 crc kubenswrapper[4775]: I0321 04:47:59.604515 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:00 crc kubenswrapper[4775]: I0321 04:48:00.604840 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:01 crc kubenswrapper[4775]: I0321 04:48:01.604936 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:02 crc kubenswrapper[4775]: I0321 04:48:02.315384 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:48:02 crc kubenswrapper[4775]: E0321 04:48:02.316444 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 04:48:02 crc kubenswrapper[4775]: I0321 04:48:02.317354 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:02 crc kubenswrapper[4775]: I0321 04:48:02.317508 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:02 crc kubenswrapper[4775]: I0321 04:48:02.317535 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:02 crc kubenswrapper[4775]: I0321 04:48:02.317576 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 04:48:02 crc kubenswrapper[4775]: E0321 04:48:02.321710 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 04:48:02 crc kubenswrapper[4775]: I0321 04:48:02.395049 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:48:02 crc kubenswrapper[4775]: I0321 04:48:02.395363 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:48:02 crc kubenswrapper[4775]: I0321 04:48:02.397034 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:02 crc kubenswrapper[4775]: I0321 04:48:02.397099 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:02 crc kubenswrapper[4775]: I0321 04:48:02.397145 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:02 crc kubenswrapper[4775]: I0321 04:48:02.604446 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:03 crc kubenswrapper[4775]: I0321 04:48:03.603847 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:03 crc kubenswrapper[4775]: W0321 04:48:03.699067 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 21 04:48:03 crc kubenswrapper[4775]: E0321 04:48:03.699181 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 21 04:48:04 crc kubenswrapper[4775]: I0321 04:48:04.604973 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:05 crc kubenswrapper[4775]: W0321 04:48:05.394185 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:05 crc kubenswrapper[4775]: E0321 04:48:05.394437 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 21 04:48:05 crc kubenswrapper[4775]: I0321 04:48:05.395224 4775 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 21 04:48:05 crc kubenswrapper[4775]: I0321 04:48:05.395279 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 21 04:48:05 crc kubenswrapper[4775]: E0321 04:48:05.399568 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189ec1dc5b06d854\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 21 04:48:05 crc kubenswrapper[4775]: &Event{ObjectMeta:{kube-controller-manager-crc.189ec1dc5b06d854 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 21 04:48:05 crc kubenswrapper[4775]: body: Mar 21 04:48:05 crc kubenswrapper[4775]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:35.394695252 +0000 UTC m=+8.371158876,LastTimestamp:2026-03-21 04:48:05.395264485 +0000 UTC m=+38.371728109,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 21 04:48:05 crc kubenswrapper[4775]: > Mar 21 04:48:05 crc kubenswrapper[4775]: E0321 04:48:05.404506 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189ec1dc5b07b0b4\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ec1dc5b07b0b4 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:35.394750644 +0000 UTC m=+8.371214268,LastTimestamp:2026-03-21 04:48:05.395302146 +0000 UTC m=+38.371765770,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:48:05 crc kubenswrapper[4775]: I0321 04:48:05.603817 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:06 crc kubenswrapper[4775]: I0321 04:48:06.612697 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:06 crc kubenswrapper[4775]: W0321 04:48:06.876411 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 21 04:48:06 crc kubenswrapper[4775]: E0321 04:48:06.876471 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 21 04:48:07 crc kubenswrapper[4775]: I0321 04:48:07.606217 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:07 crc kubenswrapper[4775]: E0321 04:48:07.733238 4775 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 04:48:08 crc kubenswrapper[4775]: I0321 04:48:08.604887 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:08 crc kubenswrapper[4775]: I0321 04:48:08.661312 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:48:08 crc kubenswrapper[4775]: I0321 04:48:08.662691 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:08 crc kubenswrapper[4775]: I0321 04:48:08.662744 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:08 crc kubenswrapper[4775]: I0321 04:48:08.662755 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:08 crc kubenswrapper[4775]: I0321 04:48:08.663396 4775 scope.go:117] "RemoveContainer" containerID="a3ec38c653682ae3ceeecec7bf83cf11591c8de9989f8ce7f4b2ce4297b763cc" Mar 21 04:48:08 crc kubenswrapper[4775]: I0321 04:48:08.843239 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 21 04:48:08 crc kubenswrapper[4775]: I0321 04:48:08.844976 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9756a73583a81b5cd100849e9bf3b7b2cbe36f2d69afc978be7f77672bb9a36b"} Mar 21 04:48:08 crc kubenswrapper[4775]: I0321 04:48:08.845146 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:48:08 crc kubenswrapper[4775]: I0321 04:48:08.845848 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:08 crc kubenswrapper[4775]: I0321 04:48:08.845886 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:08 crc kubenswrapper[4775]: I0321 04:48:08.845897 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:09 crc kubenswrapper[4775]: E0321 04:48:09.320933 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 04:48:09 crc kubenswrapper[4775]: I0321 04:48:09.321816 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:48:09 crc kubenswrapper[4775]: I0321 04:48:09.322931 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:09 crc kubenswrapper[4775]: I0321 04:48:09.322956 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:09 crc kubenswrapper[4775]: I0321 04:48:09.322967 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:09 crc kubenswrapper[4775]: I0321 04:48:09.322999 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 04:48:09 crc kubenswrapper[4775]: E0321 04:48:09.327953 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 04:48:09 crc kubenswrapper[4775]: I0321 04:48:09.606142 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:09 crc kubenswrapper[4775]: I0321 04:48:09.849071 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 21 04:48:09 crc kubenswrapper[4775]: I0321 04:48:09.849532 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 21 04:48:09 crc kubenswrapper[4775]: I0321 04:48:09.851024 4775 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9756a73583a81b5cd100849e9bf3b7b2cbe36f2d69afc978be7f77672bb9a36b" exitCode=255 Mar 21 04:48:09 crc kubenswrapper[4775]: I0321 04:48:09.851067 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9756a73583a81b5cd100849e9bf3b7b2cbe36f2d69afc978be7f77672bb9a36b"} Mar 21 04:48:09 crc kubenswrapper[4775]: I0321 04:48:09.851105 4775 scope.go:117] "RemoveContainer" containerID="a3ec38c653682ae3ceeecec7bf83cf11591c8de9989f8ce7f4b2ce4297b763cc" Mar 21 04:48:09 crc kubenswrapper[4775]: I0321 04:48:09.851254 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:48:09 crc kubenswrapper[4775]: I0321 04:48:09.852323 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:09 crc kubenswrapper[4775]: I0321 04:48:09.852377 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:09 crc kubenswrapper[4775]: I0321 04:48:09.852391 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:09 crc kubenswrapper[4775]: I0321 04:48:09.853061 4775 scope.go:117] "RemoveContainer" containerID="9756a73583a81b5cd100849e9bf3b7b2cbe36f2d69afc978be7f77672bb9a36b" Mar 21 04:48:09 crc kubenswrapper[4775]: E0321 04:48:09.853273 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 04:48:10 crc kubenswrapper[4775]: I0321 04:48:10.605832 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:10 crc kubenswrapper[4775]: I0321 04:48:10.855238 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 21 04:48:11 crc kubenswrapper[4775]: I0321 04:48:11.605577 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:12 crc kubenswrapper[4775]: I0321 04:48:12.604992 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:13 crc kubenswrapper[4775]: I0321 04:48:13.133029 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:48:13 crc kubenswrapper[4775]: I0321 04:48:13.133243 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:48:13 crc kubenswrapper[4775]: I0321 04:48:13.134257 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:13 crc kubenswrapper[4775]: I0321 04:48:13.134290 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:13 crc kubenswrapper[4775]: I0321 04:48:13.134301 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:13 crc kubenswrapper[4775]: I0321 04:48:13.134858 4775 scope.go:117] "RemoveContainer" containerID="9756a73583a81b5cd100849e9bf3b7b2cbe36f2d69afc978be7f77672bb9a36b" Mar 21 04:48:13 crc kubenswrapper[4775]: E0321 04:48:13.135027 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 04:48:14 crc kubenswrapper[4775]: I0321 04:48:14.479557 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:14 crc kubenswrapper[4775]: I0321 04:48:14.601250 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 04:48:14 crc kubenswrapper[4775]: I0321 04:48:14.601456 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:48:14 crc kubenswrapper[4775]: I0321 04:48:14.603111 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:14 crc kubenswrapper[4775]: I0321 04:48:14.603199 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:14 crc kubenswrapper[4775]: I0321 04:48:14.603217 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:14 crc kubenswrapper[4775]: I0321 04:48:14.603771 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:15 crc kubenswrapper[4775]: I0321 04:48:15.395861 4775 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 21 04:48:15 crc kubenswrapper[4775]: I0321 04:48:15.395953 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 21 04:48:15 crc kubenswrapper[4775]: E0321 04:48:15.401418 4775 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189ec1dc5b06d854\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 21 04:48:15 crc kubenswrapper[4775]: &Event{ObjectMeta:{kube-controller-manager-crc.189ec1dc5b06d854 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 21 04:48:15 crc kubenswrapper[4775]: body: Mar 21 04:48:15 crc kubenswrapper[4775]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:47:35.394695252 +0000 UTC m=+8.371158876,LastTimestamp:2026-03-21 04:48:15.395933439 +0000 UTC m=+48.372397063,Count:5,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 21 04:48:15 crc kubenswrapper[4775]: > Mar 21 04:48:15 crc kubenswrapper[4775]: I0321 04:48:15.604758 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:16 crc kubenswrapper[4775]: E0321 04:48:16.326083 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 04:48:16 crc kubenswrapper[4775]: I0321 04:48:16.328043 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:48:16 crc kubenswrapper[4775]: I0321 04:48:16.329350 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:16 crc kubenswrapper[4775]: I0321 04:48:16.329393 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:16 crc kubenswrapper[4775]: I0321 04:48:16.329405 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:16 crc kubenswrapper[4775]: I0321 04:48:16.329444 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 04:48:16 crc kubenswrapper[4775]: E0321 04:48:16.334435 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 04:48:16 crc kubenswrapper[4775]: I0321 04:48:16.604455 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:17 crc kubenswrapper[4775]: I0321 04:48:17.604343 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:17 crc kubenswrapper[4775]: E0321 04:48:17.734161 4775 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 04:48:18 crc kubenswrapper[4775]: W0321 04:48:18.499460 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 21 04:48:18 crc kubenswrapper[4775]: E0321 04:48:18.499572 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 21 04:48:18 crc kubenswrapper[4775]: I0321 04:48:18.605554 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:19 crc kubenswrapper[4775]: I0321 04:48:19.606686 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:19 crc kubenswrapper[4775]: I0321 04:48:19.703372 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:48:19 crc kubenswrapper[4775]: I0321 04:48:19.703652 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:48:19 crc kubenswrapper[4775]: I0321 04:48:19.704723 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:19 crc kubenswrapper[4775]: I0321 04:48:19.704777 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:19 crc kubenswrapper[4775]: I0321 04:48:19.704788 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:19 crc kubenswrapper[4775]: I0321 04:48:19.705390 4775 scope.go:117] "RemoveContainer" containerID="9756a73583a81b5cd100849e9bf3b7b2cbe36f2d69afc978be7f77672bb9a36b" Mar 21 04:48:19 crc kubenswrapper[4775]: E0321 04:48:19.705600 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 04:48:20 crc kubenswrapper[4775]: I0321 04:48:20.606309 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:21 crc kubenswrapper[4775]: I0321 04:48:21.604533 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:22 crc kubenswrapper[4775]: I0321 04:48:22.606210 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:23 crc kubenswrapper[4775]: E0321 04:48:23.331724 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 04:48:23 crc kubenswrapper[4775]: I0321 04:48:23.334808 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:48:23 crc kubenswrapper[4775]: I0321 04:48:23.336159 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:23 crc kubenswrapper[4775]: I0321 04:48:23.336198 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:23 crc kubenswrapper[4775]: I0321 04:48:23.336210 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:23 crc kubenswrapper[4775]: I0321 04:48:23.336242 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 04:48:23 crc kubenswrapper[4775]: E0321 04:48:23.340838 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 04:48:23 crc kubenswrapper[4775]: I0321 04:48:23.605046 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:24 crc kubenswrapper[4775]: I0321 04:48:24.020623 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:48:24 crc kubenswrapper[4775]: I0321 04:48:24.020796 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:48:24 crc kubenswrapper[4775]: I0321 04:48:24.021928 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:24 crc kubenswrapper[4775]: I0321 04:48:24.021965 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:24 crc kubenswrapper[4775]: I0321 04:48:24.021980 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:24 crc kubenswrapper[4775]: I0321 04:48:24.030050 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:48:24 crc kubenswrapper[4775]: I0321 04:48:24.500851 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:48:24 crc kubenswrapper[4775]: I0321 04:48:24.502649 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:24 crc kubenswrapper[4775]: I0321 04:48:24.502701 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:24 crc kubenswrapper[4775]: I0321 04:48:24.502716 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:24 crc kubenswrapper[4775]: I0321 04:48:24.606522 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:25 crc kubenswrapper[4775]: I0321 04:48:25.604663 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:26 crc kubenswrapper[4775]: I0321 04:48:26.606049 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:27 crc kubenswrapper[4775]: I0321 04:48:27.604100 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:27 crc kubenswrapper[4775]: E0321 04:48:27.734928 4775 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 04:48:28 crc kubenswrapper[4775]: I0321 04:48:28.603948 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:29 crc kubenswrapper[4775]: I0321 04:48:29.604368 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:30 crc kubenswrapper[4775]: E0321 04:48:30.337587 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 04:48:30 crc kubenswrapper[4775]: I0321 04:48:30.341728 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:48:30 crc kubenswrapper[4775]: I0321 04:48:30.343087 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:30 crc kubenswrapper[4775]: I0321 04:48:30.343131 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:30 crc kubenswrapper[4775]: I0321 04:48:30.343144 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:30 crc kubenswrapper[4775]: I0321 04:48:30.343173 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 04:48:30 crc kubenswrapper[4775]: E0321 04:48:30.347421 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 04:48:30 crc kubenswrapper[4775]: I0321 04:48:30.607360 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:31 crc kubenswrapper[4775]: I0321 04:48:31.604555 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 04:48:32 crc kubenswrapper[4775]: I0321 04:48:32.113865 4775 csr.go:261] certificate signing request csr-mkg6s is approved, waiting to be issued Mar 21 04:48:32 crc kubenswrapper[4775]: I0321 04:48:32.122827 4775 csr.go:257] certificate signing request csr-mkg6s is issued Mar 21 04:48:32 crc kubenswrapper[4775]: I0321 04:48:32.150515 4775 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 21 04:48:32 crc kubenswrapper[4775]: I0321 04:48:32.461312 4775 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 21 04:48:32 crc kubenswrapper[4775]: I0321 04:48:32.660611 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:48:32 crc kubenswrapper[4775]: I0321 04:48:32.661577 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:32 crc kubenswrapper[4775]: I0321 04:48:32.661608 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:32 crc kubenswrapper[4775]: I0321 04:48:32.661616 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:32 crc kubenswrapper[4775]: I0321 04:48:32.662216 4775 scope.go:117] "RemoveContainer" containerID="9756a73583a81b5cd100849e9bf3b7b2cbe36f2d69afc978be7f77672bb9a36b" Mar 21 04:48:33 crc kubenswrapper[4775]: I0321 04:48:33.123990 4775 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-28 22:05:43.809376382 +0000 UTC Mar 21 04:48:33 crc kubenswrapper[4775]: I0321 04:48:33.124358 4775 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6065h17m10.685022258s for next certificate rotation Mar 21 04:48:33 crc kubenswrapper[4775]: I0321 04:48:33.521226 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 21 04:48:33 crc kubenswrapper[4775]: I0321 04:48:33.521934 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 21 04:48:33 crc kubenswrapper[4775]: I0321 04:48:33.523657 4775 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4" exitCode=255 Mar 21 04:48:33 crc kubenswrapper[4775]: I0321 04:48:33.523705 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4"} Mar 21 04:48:33 crc kubenswrapper[4775]: I0321 04:48:33.523743 4775 scope.go:117] "RemoveContainer" containerID="9756a73583a81b5cd100849e9bf3b7b2cbe36f2d69afc978be7f77672bb9a36b" Mar 21 04:48:33 crc kubenswrapper[4775]: I0321 04:48:33.523945 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:48:33 crc kubenswrapper[4775]: I0321 04:48:33.525663 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:33 crc kubenswrapper[4775]: I0321 04:48:33.525687 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:33 crc kubenswrapper[4775]: I0321 04:48:33.525697 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:33 crc kubenswrapper[4775]: I0321 04:48:33.526244 4775 scope.go:117] "RemoveContainer" containerID="85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4" Mar 21 04:48:33 crc kubenswrapper[4775]: E0321 04:48:33.526446 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 04:48:34 crc kubenswrapper[4775]: I0321 04:48:34.527601 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.347866 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.349096 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.349162 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.349175 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.349296 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.356491 4775 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.356772 4775 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 21 04:48:37 crc kubenswrapper[4775]: E0321 04:48:37.356788 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.359812 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.359854 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.359866 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.359880 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.359891 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:37Z","lastTransitionTime":"2026-03-21T04:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:37 crc kubenswrapper[4775]: E0321 04:48:37.371436 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.377665 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.377714 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.377729 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.377749 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.377763 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:37Z","lastTransitionTime":"2026-03-21T04:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:37 crc kubenswrapper[4775]: E0321 04:48:37.386748 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.392929 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.392965 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.392976 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.392992 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.393003 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:37Z","lastTransitionTime":"2026-03-21T04:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:37 crc kubenswrapper[4775]: E0321 04:48:37.402375 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.412379 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.412439 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.412452 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.412471 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:37 crc kubenswrapper[4775]: I0321 04:48:37.412487 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:37Z","lastTransitionTime":"2026-03-21T04:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:37 crc kubenswrapper[4775]: E0321 04:48:37.422003 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:37 crc kubenswrapper[4775]: E0321 04:48:37.422136 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 21 04:48:37 crc kubenswrapper[4775]: E0321 04:48:37.422160 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:37 crc kubenswrapper[4775]: E0321 04:48:37.522655 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:37 crc kubenswrapper[4775]: E0321 04:48:37.623268 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:37 crc kubenswrapper[4775]: E0321 04:48:37.723920 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:37 crc kubenswrapper[4775]: E0321 04:48:37.736389 4775 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 04:48:37 crc kubenswrapper[4775]: E0321 04:48:37.824988 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:37 crc kubenswrapper[4775]: E0321 04:48:37.925411 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:38 crc kubenswrapper[4775]: E0321 04:48:38.026430 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:38 crc kubenswrapper[4775]: E0321 04:48:38.126611 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:38 crc kubenswrapper[4775]: E0321 04:48:38.227565 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:38 crc kubenswrapper[4775]: E0321 04:48:38.328895 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:38 crc kubenswrapper[4775]: E0321 04:48:38.429860 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:38 crc kubenswrapper[4775]: E0321 04:48:38.530761 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:38 crc kubenswrapper[4775]: E0321 04:48:38.631470 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:38 crc kubenswrapper[4775]: E0321 04:48:38.732188 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:38 crc kubenswrapper[4775]: E0321 04:48:38.832764 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:38 crc kubenswrapper[4775]: E0321 04:48:38.933500 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:39 crc kubenswrapper[4775]: E0321 04:48:39.033873 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:39 crc kubenswrapper[4775]: E0321 04:48:39.134248 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:39 crc kubenswrapper[4775]: E0321 04:48:39.235338 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:39 crc kubenswrapper[4775]: E0321 04:48:39.335715 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:39 crc kubenswrapper[4775]: E0321 04:48:39.435849 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:39 crc kubenswrapper[4775]: E0321 04:48:39.536404 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:39 crc kubenswrapper[4775]: E0321 04:48:39.636994 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:39 crc kubenswrapper[4775]: I0321 04:48:39.702811 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:48:39 crc kubenswrapper[4775]: I0321 04:48:39.702962 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:48:39 crc kubenswrapper[4775]: I0321 04:48:39.703948 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:39 crc kubenswrapper[4775]: I0321 04:48:39.703987 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:39 crc kubenswrapper[4775]: I0321 04:48:39.704002 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:39 crc kubenswrapper[4775]: I0321 04:48:39.704550 4775 scope.go:117] "RemoveContainer" containerID="85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4" Mar 21 04:48:39 crc kubenswrapper[4775]: E0321 04:48:39.704725 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 04:48:39 crc kubenswrapper[4775]: E0321 04:48:39.737745 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:39 crc kubenswrapper[4775]: E0321 04:48:39.838095 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:39 crc kubenswrapper[4775]: E0321 04:48:39.938924 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:40 crc kubenswrapper[4775]: E0321 04:48:40.039263 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:40 crc kubenswrapper[4775]: I0321 04:48:40.069040 4775 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 21 04:48:40 crc kubenswrapper[4775]: E0321 04:48:40.139541 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:40 crc kubenswrapper[4775]: E0321 04:48:40.239664 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:40 crc kubenswrapper[4775]: E0321 04:48:40.340525 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:40 crc kubenswrapper[4775]: E0321 04:48:40.441135 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:40 crc kubenswrapper[4775]: E0321 04:48:40.541226 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:40 crc kubenswrapper[4775]: E0321 04:48:40.641900 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:40 crc kubenswrapper[4775]: E0321 04:48:40.742538 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:40 crc kubenswrapper[4775]: E0321 04:48:40.843650 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:40 crc kubenswrapper[4775]: E0321 04:48:40.944060 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:40 crc kubenswrapper[4775]: I0321 04:48:40.981169 4775 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 21 04:48:41 crc kubenswrapper[4775]: E0321 04:48:41.044785 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:41 crc kubenswrapper[4775]: E0321 04:48:41.145242 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:41 crc kubenswrapper[4775]: E0321 04:48:41.246345 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:41 crc kubenswrapper[4775]: E0321 04:48:41.346849 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:41 crc kubenswrapper[4775]: E0321 04:48:41.447865 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:41 crc kubenswrapper[4775]: E0321 04:48:41.548502 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:41 crc kubenswrapper[4775]: E0321 04:48:41.648797 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:41 crc kubenswrapper[4775]: E0321 04:48:41.749170 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:41 crc kubenswrapper[4775]: E0321 04:48:41.849424 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:41 crc kubenswrapper[4775]: E0321 04:48:41.950548 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:42 crc kubenswrapper[4775]: E0321 04:48:42.051185 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:42 crc kubenswrapper[4775]: E0321 04:48:42.151811 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:42 crc kubenswrapper[4775]: E0321 04:48:42.252663 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:42 crc kubenswrapper[4775]: E0321 04:48:42.353595 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:42 crc kubenswrapper[4775]: E0321 04:48:42.453895 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:42 crc kubenswrapper[4775]: E0321 04:48:42.554570 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:42 crc kubenswrapper[4775]: E0321 04:48:42.654841 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:42 crc kubenswrapper[4775]: E0321 04:48:42.755147 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:42 crc kubenswrapper[4775]: E0321 04:48:42.855597 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:42 crc kubenswrapper[4775]: E0321 04:48:42.955974 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:43 crc kubenswrapper[4775]: E0321 04:48:43.056152 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:43 crc kubenswrapper[4775]: I0321 04:48:43.133581 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:48:43 crc kubenswrapper[4775]: I0321 04:48:43.134110 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:48:43 crc kubenswrapper[4775]: I0321 04:48:43.135444 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:43 crc kubenswrapper[4775]: I0321 04:48:43.135481 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:43 crc kubenswrapper[4775]: I0321 04:48:43.135494 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:43 crc kubenswrapper[4775]: I0321 04:48:43.136167 4775 scope.go:117] "RemoveContainer" containerID="85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4" Mar 21 04:48:43 crc kubenswrapper[4775]: E0321 04:48:43.136383 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 04:48:43 crc kubenswrapper[4775]: E0321 04:48:43.156625 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:43 crc kubenswrapper[4775]: E0321 04:48:43.257410 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:43 crc kubenswrapper[4775]: E0321 04:48:43.357829 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:43 crc kubenswrapper[4775]: E0321 04:48:43.458694 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:43 crc kubenswrapper[4775]: E0321 04:48:43.559102 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:43 crc kubenswrapper[4775]: E0321 04:48:43.660331 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:43 crc kubenswrapper[4775]: E0321 04:48:43.761228 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:43 crc kubenswrapper[4775]: E0321 04:48:43.861445 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:43 crc kubenswrapper[4775]: E0321 04:48:43.962225 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:44 crc kubenswrapper[4775]: E0321 04:48:44.063432 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:44 crc kubenswrapper[4775]: E0321 04:48:44.164523 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:44 crc kubenswrapper[4775]: E0321 04:48:44.265291 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:44 crc kubenswrapper[4775]: E0321 04:48:44.366034 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:44 crc kubenswrapper[4775]: E0321 04:48:44.466966 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:44 crc kubenswrapper[4775]: E0321 04:48:44.568178 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:44 crc kubenswrapper[4775]: E0321 04:48:44.668539 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:44 crc kubenswrapper[4775]: E0321 04:48:44.769379 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:44 crc kubenswrapper[4775]: E0321 04:48:44.870196 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:44 crc kubenswrapper[4775]: E0321 04:48:44.971027 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:45 crc kubenswrapper[4775]: E0321 04:48:45.072202 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:45 crc kubenswrapper[4775]: E0321 04:48:45.172586 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:45 crc kubenswrapper[4775]: E0321 04:48:45.273416 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:45 crc kubenswrapper[4775]: E0321 04:48:45.374417 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:45 crc kubenswrapper[4775]: E0321 04:48:45.475585 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:45 crc kubenswrapper[4775]: E0321 04:48:45.576060 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:45 crc kubenswrapper[4775]: I0321 04:48:45.660778 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 04:48:45 crc kubenswrapper[4775]: I0321 04:48:45.662010 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:45 crc kubenswrapper[4775]: I0321 04:48:45.662057 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:45 crc kubenswrapper[4775]: I0321 04:48:45.662082 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:45 crc kubenswrapper[4775]: E0321 04:48:45.676610 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:45 crc kubenswrapper[4775]: E0321 04:48:45.776721 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:45 crc kubenswrapper[4775]: E0321 04:48:45.877796 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:45 crc kubenswrapper[4775]: E0321 04:48:45.978930 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:46 crc kubenswrapper[4775]: E0321 04:48:46.079611 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:46 crc kubenswrapper[4775]: E0321 04:48:46.180409 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:46 crc kubenswrapper[4775]: E0321 04:48:46.281287 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:46 crc kubenswrapper[4775]: E0321 04:48:46.382284 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:46 crc kubenswrapper[4775]: E0321 04:48:46.483367 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:46 crc kubenswrapper[4775]: E0321 04:48:46.584352 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:46 crc kubenswrapper[4775]: E0321 04:48:46.685461 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:46 crc kubenswrapper[4775]: E0321 04:48:46.785888 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:46 crc kubenswrapper[4775]: E0321 04:48:46.887019 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:46 crc kubenswrapper[4775]: E0321 04:48:46.987401 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:47 crc kubenswrapper[4775]: E0321 04:48:47.087595 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:47 crc kubenswrapper[4775]: E0321 04:48:47.188643 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:47 crc kubenswrapper[4775]: E0321 04:48:47.288928 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:47 crc kubenswrapper[4775]: E0321 04:48:47.389789 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:47 crc kubenswrapper[4775]: E0321 04:48:47.490665 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:47 crc kubenswrapper[4775]: E0321 04:48:47.590945 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:47 crc kubenswrapper[4775]: E0321 04:48:47.691599 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:47 crc kubenswrapper[4775]: E0321 04:48:47.737315 4775 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 04:48:47 crc kubenswrapper[4775]: E0321 04:48:47.743936 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 21 04:48:47 crc kubenswrapper[4775]: I0321 04:48:47.748415 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:47 crc kubenswrapper[4775]: I0321 04:48:47.748454 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:47 crc kubenswrapper[4775]: I0321 04:48:47.748463 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:47 crc kubenswrapper[4775]: I0321 04:48:47.748476 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:47 crc kubenswrapper[4775]: I0321 04:48:47.748485 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:47Z","lastTransitionTime":"2026-03-21T04:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:47 crc kubenswrapper[4775]: E0321 04:48:47.759554 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:47 crc kubenswrapper[4775]: I0321 04:48:47.763774 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:47 crc kubenswrapper[4775]: I0321 04:48:47.763813 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:47 crc kubenswrapper[4775]: I0321 04:48:47.763827 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:47 crc kubenswrapper[4775]: I0321 04:48:47.763844 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:47 crc kubenswrapper[4775]: I0321 04:48:47.763855 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:47Z","lastTransitionTime":"2026-03-21T04:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:47 crc kubenswrapper[4775]: E0321 04:48:47.773735 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:47 crc kubenswrapper[4775]: I0321 04:48:47.791737 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:47 crc kubenswrapper[4775]: I0321 04:48:47.791783 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:47 crc kubenswrapper[4775]: I0321 04:48:47.791791 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:47 crc kubenswrapper[4775]: I0321 04:48:47.791806 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:47 crc kubenswrapper[4775]: I0321 04:48:47.791815 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:47Z","lastTransitionTime":"2026-03-21T04:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:47 crc kubenswrapper[4775]: E0321 04:48:47.800679 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:47 crc kubenswrapper[4775]: I0321 04:48:47.804204 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:47 crc kubenswrapper[4775]: I0321 04:48:47.804342 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:47 crc kubenswrapper[4775]: I0321 04:48:47.804433 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:47 crc kubenswrapper[4775]: I0321 04:48:47.804566 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:47 crc kubenswrapper[4775]: I0321 04:48:47.804659 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:47Z","lastTransitionTime":"2026-03-21T04:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:47 crc kubenswrapper[4775]: E0321 04:48:47.814398 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:47 crc kubenswrapper[4775]: E0321 04:48:47.814513 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 21 04:48:47 crc kubenswrapper[4775]: E0321 04:48:47.814535 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:47 crc kubenswrapper[4775]: E0321 04:48:47.915204 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:48 crc kubenswrapper[4775]: E0321 04:48:48.016411 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:48 crc kubenswrapper[4775]: E0321 04:48:48.117154 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:48 crc kubenswrapper[4775]: E0321 04:48:48.218318 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:48 crc kubenswrapper[4775]: E0321 04:48:48.318783 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:48 crc kubenswrapper[4775]: E0321 04:48:48.419828 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:48 crc kubenswrapper[4775]: E0321 04:48:48.520593 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:48 crc kubenswrapper[4775]: E0321 04:48:48.621559 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:48 crc kubenswrapper[4775]: E0321 04:48:48.722149 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:48 crc kubenswrapper[4775]: E0321 04:48:48.822762 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:48 crc kubenswrapper[4775]: E0321 04:48:48.923700 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:49 crc kubenswrapper[4775]: E0321 04:48:49.024053 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:49 crc kubenswrapper[4775]: E0321 04:48:49.124744 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:49 crc kubenswrapper[4775]: E0321 04:48:49.225331 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:49 crc kubenswrapper[4775]: E0321 04:48:49.326247 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:49 crc kubenswrapper[4775]: E0321 04:48:49.427232 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:49 crc kubenswrapper[4775]: E0321 04:48:49.528045 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:49 crc kubenswrapper[4775]: E0321 04:48:49.629416 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:49 crc kubenswrapper[4775]: E0321 04:48:49.730217 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:49 crc kubenswrapper[4775]: E0321 04:48:49.830950 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:49 crc kubenswrapper[4775]: E0321 04:48:49.931608 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:50 crc kubenswrapper[4775]: E0321 04:48:50.031894 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:50 crc kubenswrapper[4775]: E0321 04:48:50.132440 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:50 crc kubenswrapper[4775]: E0321 04:48:50.232807 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:50 crc kubenswrapper[4775]: E0321 04:48:50.333583 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:50 crc kubenswrapper[4775]: E0321 04:48:50.434610 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:50 crc kubenswrapper[4775]: E0321 04:48:50.535427 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:50 crc kubenswrapper[4775]: E0321 04:48:50.635980 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:50 crc kubenswrapper[4775]: E0321 04:48:50.737106 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:50 crc kubenswrapper[4775]: E0321 04:48:50.837566 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:50 crc kubenswrapper[4775]: E0321 04:48:50.938578 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:51 crc kubenswrapper[4775]: E0321 04:48:51.039752 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:51 crc kubenswrapper[4775]: E0321 04:48:51.140402 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:51 crc kubenswrapper[4775]: E0321 04:48:51.240744 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:51 crc kubenswrapper[4775]: E0321 04:48:51.341026 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:51 crc kubenswrapper[4775]: E0321 04:48:51.441871 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:51 crc kubenswrapper[4775]: E0321 04:48:51.542502 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.571107 4775 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.645777 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.645829 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.645846 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.645864 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.645875 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:51Z","lastTransitionTime":"2026-03-21T04:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.748182 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.748221 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.748231 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.748245 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.748254 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:51Z","lastTransitionTime":"2026-03-21T04:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.850569 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.850611 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.850622 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.850638 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.850649 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:51Z","lastTransitionTime":"2026-03-21T04:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.953665 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.953696 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.953704 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.953717 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:51 crc kubenswrapper[4775]: I0321 04:48:51.953725 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:51Z","lastTransitionTime":"2026-03-21T04:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.056236 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.056273 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.056283 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.056297 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.056305 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:52Z","lastTransitionTime":"2026-03-21T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.158448 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.158476 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.158485 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.158498 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.158506 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:52Z","lastTransitionTime":"2026-03-21T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.261102 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.261175 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.261192 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.261210 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.261237 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:52Z","lastTransitionTime":"2026-03-21T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.363575 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.363604 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.363617 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.363632 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.363641 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:52Z","lastTransitionTime":"2026-03-21T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.466207 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.466257 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.466273 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.466290 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.466300 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:52Z","lastTransitionTime":"2026-03-21T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.499068 4775 apiserver.go:52] "Watching apiserver" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.505059 4775 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.505355 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.505665 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.505772 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.505908 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.505943 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.506058 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.506104 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.506185 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.506259 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.506385 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.508555 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.508926 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.509145 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.510255 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.510399 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.510458 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.510399 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.510814 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.510833 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.533991 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.549673 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.562030 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.568942 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.568982 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.568992 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.569008 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.569021 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:52Z","lastTransitionTime":"2026-03-21T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.575840 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.590139 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.603438 4775 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.606605 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.624539 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.639625 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.650606 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.661135 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.671198 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.671234 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.671242 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.671256 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.671265 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:52Z","lastTransitionTime":"2026-03-21T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.671314 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690552 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690605 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690624 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690641 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690665 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690683 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690705 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690726 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690742 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690761 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690783 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690803 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690824 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690848 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690869 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690889 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690908 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690940 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690961 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690978 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.690994 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.691012 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.691033 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.691051 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.691069 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.691088 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.691129 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.691150 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.691150 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.691167 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.691187 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.691173 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.691207 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.691232 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.691251 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.691374 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.694030 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.696464 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.691438 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.691986 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.692108 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.691180 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.692141 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.692341 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.692359 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.692464 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.692346 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.692636 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.692813 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.692829 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.696506 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.696983 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.697020 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.697047 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.697066 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.697090 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.697133 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.697160 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.697188 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.697208 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.697230 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.697248 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.697276 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.697330 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.697338 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.697399 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.697431 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.697463 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.697489 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.692930 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.693093 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.693103 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.697643 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.697866 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.693156 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.693188 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.693246 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.693260 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.693352 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.693593 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.693657 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.693718 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.693959 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.693982 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.693989 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.694282 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.694446 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.695064 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.698146 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.698205 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.699168 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.699516 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.697883 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.699680 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.699710 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.699737 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.699745 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.699768 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.699805 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.699769 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.699894 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.699930 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.699962 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.699990 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.700029 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.700058 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.700085 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.700142 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.700172 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.700199 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.700226 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.699961 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.700253 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.700422 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.700535 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.700632 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.700919 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.700328 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701188 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701216 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701234 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701252 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701270 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701287 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701305 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701321 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701339 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701356 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701374 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701394 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701411 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701429 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701445 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701462 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701479 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701497 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701513 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701530 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701547 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701563 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701579 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701596 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701629 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701657 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701678 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701700 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701725 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701752 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701773 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701793 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701809 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701825 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701843 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701859 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701880 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701895 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701913 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701929 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701944 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701961 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701983 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.701999 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.702015 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.702031 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.702048 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.702063 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.702369 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.702388 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.702415 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.702441 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.702581 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.702601 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.702902 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.702468 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703258 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703297 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703326 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703354 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703379 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703405 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703429 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703452 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703480 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703502 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703526 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703549 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703576 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703632 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703659 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703702 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703796 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703817 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703835 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703851 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703872 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703892 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703910 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703926 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703944 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703964 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703987 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.704011 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.704030 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.704052 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.704079 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.704099 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.704325 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703439 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703557 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703620 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703619 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703750 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703870 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.703968 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.704306 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.704322 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:48:53.204296336 +0000 UTC m=+86.180759960 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.705391 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.705439 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.705468 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.705500 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.705523 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.705551 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.705578 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.705597 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.705614 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.705632 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.705650 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.705667 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.705685 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.705752 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.705772 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.705789 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.706069 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.706094 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.706111 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.706245 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.706273 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.706291 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.706355 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.706372 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.706389 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.706407 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.707901 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.707945 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708003 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708091 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708134 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708185 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708210 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708240 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708268 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708290 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708318 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708346 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708371 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708673 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708717 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708748 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708775 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708803 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708880 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708907 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708937 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708966 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708999 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.709025 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.709055 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.709083 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.709112 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710328 4775 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710357 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710374 4775 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710389 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710404 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710417 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710432 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710461 4775 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710477 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710491 4775 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710504 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710518 4775 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710531 4775 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710543 4775 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710556 4775 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710571 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710585 4775 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710599 4775 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713176 4775 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713386 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713404 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713419 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713619 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713653 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713673 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713691 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713706 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713728 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713742 4775 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713757 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713772 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713784 4775 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713800 4775 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713814 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713825 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713837 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713850 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713864 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713876 4775 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713889 4775 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713903 4775 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713914 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713927 4775 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713940 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713952 4775 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713977 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713990 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714005 4775 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714017 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714033 4775 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714046 4775 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714058 4775 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714068 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714078 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714091 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714106 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714135 4775 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714146 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714159 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.711786 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714793 4775 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.706149 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.706593 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.705094 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.705098 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.706927 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.706955 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.706995 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.707378 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.707531 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.707548 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.707745 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.707884 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.707969 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708365 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708386 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.708614 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.709484 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.709637 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.709704 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.709779 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.709785 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710080 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710186 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710324 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.709801 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710549 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710595 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710736 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.710858 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.711020 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.711008 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.711490 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.711634 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.711789 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.705090 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.711848 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.712425 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.712460 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.712617 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.712662 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.712747 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.712766 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.712966 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.712959 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713141 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713176 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713248 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713297 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713291 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713452 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713520 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713611 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713636 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713649 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.713906 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714154 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714193 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714221 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714248 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.714333 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.715407 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.715460 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714448 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714619 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714685 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714686 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.714866 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.715841 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.716067 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.716094 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.716530 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.716789 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.716833 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.716541 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 04:48:53.216518781 +0000 UTC m=+86.192982405 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.717925 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 04:48:53.21787165 +0000 UTC m=+86.194335314 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.718020 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.717128 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.717134 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.717190 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.717234 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.717404 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.717444 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.717443 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.717591 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.717553 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.718650 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.719000 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.719260 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.719289 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.722841 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.724850 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.726748 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.726777 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.726793 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.726859 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-21 04:48:53.226838103 +0000 UTC m=+86.203301917 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.729724 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.730584 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.730986 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.731406 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.731476 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.731515 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.731633 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.731656 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.731724 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-21 04:48:53.23170075 +0000 UTC m=+86.208164554 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.732082 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.732181 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.732952 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.733189 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.734354 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.735696 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.738245 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.739584 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.739605 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.740536 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.741044 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.741070 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.741104 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.741050 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.741474 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.741589 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.741693 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.741824 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.742225 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.743370 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.743544 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.743759 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.743979 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.744903 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.744898 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.745082 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.745083 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.745156 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.745315 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.745405 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.745484 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.745636 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.745681 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.745965 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.746203 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.746270 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.746669 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.746741 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.746801 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.746867 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.746922 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.746972 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.748190 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.748600 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.749071 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.749280 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.761353 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.768688 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.769211 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.770552 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.773290 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.773317 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.773343 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.773362 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.773378 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:52Z","lastTransitionTime":"2026-03-21T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.814693 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.814771 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.814784 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.814846 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.814861 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.814888 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.814924 4775 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.814938 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.814947 4775 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.814956 4775 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.814965 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.814976 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.814996 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815008 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815020 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815030 4775 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815041 4775 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815049 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815057 4775 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815065 4775 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815073 4775 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815083 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815091 4775 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815100 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815108 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815146 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815160 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815169 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815178 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815186 4775 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815194 4775 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815202 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815210 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815218 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815228 4775 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815236 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815243 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815251 4775 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815261 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815269 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815277 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815285 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815293 4775 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815301 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815308 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815316 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815324 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815332 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815339 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815347 4775 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815355 4775 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815362 4775 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815370 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815377 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815385 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815393 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815401 4775 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815409 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815418 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815425 4775 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815433 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815441 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815450 4775 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815459 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815471 4775 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815483 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815492 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815501 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815512 4775 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815522 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815530 4775 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815538 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815546 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815556 4775 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815565 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815575 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815584 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815592 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815601 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815610 4775 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815619 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815628 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815637 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815680 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815688 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815697 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815706 4775 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815715 4775 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815723 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815731 4775 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815740 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815749 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815758 4775 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815769 4775 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815787 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815800 4775 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815810 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815822 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815832 4775 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815843 4775 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815855 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815865 4775 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815877 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815888 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815897 4775 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815905 4775 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815913 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815922 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815930 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815939 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815947 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815955 4775 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815963 4775 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815972 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815982 4775 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.815992 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816001 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816011 4775 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816023 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816032 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816042 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816054 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816065 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816074 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816082 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816090 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816097 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816105 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816133 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816142 4775 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816150 4775 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816159 4775 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816166 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816173 4775 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816181 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816189 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816196 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816204 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.816213 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.819412 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.825239 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.834617 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 21 04:48:52 crc kubenswrapper[4775]: W0321 04:48:52.837030 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-b1c7b006c84367f3a4790ee0ff97d47b510e4b6095d1cb4f1ca0271198041409 WatchSource:0}: Error finding container b1c7b006c84367f3a4790ee0ff97d47b510e4b6095d1cb4f1ca0271198041409: Status 404 returned error can't find the container with id b1c7b006c84367f3a4790ee0ff97d47b510e4b6095d1cb4f1ca0271198041409 Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.842273 4775 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 21 04:48:52 crc kubenswrapper[4775]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 21 04:48:52 crc kubenswrapper[4775]: if [[ -f "/env/_master" ]]; then Mar 21 04:48:52 crc kubenswrapper[4775]: set -o allexport Mar 21 04:48:52 crc kubenswrapper[4775]: source "/env/_master" Mar 21 04:48:52 crc kubenswrapper[4775]: set +o allexport Mar 21 04:48:52 crc kubenswrapper[4775]: fi Mar 21 04:48:52 crc kubenswrapper[4775]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 21 04:48:52 crc kubenswrapper[4775]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 21 04:48:52 crc kubenswrapper[4775]: ho_enable="--enable-hybrid-overlay" Mar 21 04:48:52 crc kubenswrapper[4775]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 21 04:48:52 crc kubenswrapper[4775]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 21 04:48:52 crc kubenswrapper[4775]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 21 04:48:52 crc kubenswrapper[4775]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 21 04:48:52 crc kubenswrapper[4775]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 21 04:48:52 crc kubenswrapper[4775]: --webhook-host=127.0.0.1 \ Mar 21 04:48:52 crc kubenswrapper[4775]: --webhook-port=9743 \ Mar 21 04:48:52 crc kubenswrapper[4775]: ${ho_enable} \ Mar 21 04:48:52 crc kubenswrapper[4775]: --enable-interconnect \ Mar 21 04:48:52 crc kubenswrapper[4775]: --disable-approver \ Mar 21 04:48:52 crc kubenswrapper[4775]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 21 04:48:52 crc kubenswrapper[4775]: --wait-for-kubernetes-api=200s \ Mar 21 04:48:52 crc kubenswrapper[4775]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 21 04:48:52 crc kubenswrapper[4775]: --loglevel="${LOGLEVEL}" Mar 21 04:48:52 crc kubenswrapper[4775]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 21 04:48:52 crc kubenswrapper[4775]: > logger="UnhandledError" Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.842383 4775 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 21 04:48:52 crc kubenswrapper[4775]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 21 04:48:52 crc kubenswrapper[4775]: set -o allexport Mar 21 04:48:52 crc kubenswrapper[4775]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 21 04:48:52 crc kubenswrapper[4775]: source /etc/kubernetes/apiserver-url.env Mar 21 04:48:52 crc kubenswrapper[4775]: else Mar 21 04:48:52 crc kubenswrapper[4775]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 21 04:48:52 crc kubenswrapper[4775]: exit 1 Mar 21 04:48:52 crc kubenswrapper[4775]: fi Mar 21 04:48:52 crc kubenswrapper[4775]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 21 04:48:52 crc kubenswrapper[4775]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 21 04:48:52 crc kubenswrapper[4775]: > logger="UnhandledError" Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.843497 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.844575 4775 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 21 04:48:52 crc kubenswrapper[4775]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 21 04:48:52 crc kubenswrapper[4775]: if [[ -f "/env/_master" ]]; then Mar 21 04:48:52 crc kubenswrapper[4775]: set -o allexport Mar 21 04:48:52 crc kubenswrapper[4775]: source "/env/_master" Mar 21 04:48:52 crc kubenswrapper[4775]: set +o allexport Mar 21 04:48:52 crc kubenswrapper[4775]: fi Mar 21 04:48:52 crc kubenswrapper[4775]: Mar 21 04:48:52 crc kubenswrapper[4775]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 21 04:48:52 crc kubenswrapper[4775]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 21 04:48:52 crc kubenswrapper[4775]: --disable-webhook \ Mar 21 04:48:52 crc kubenswrapper[4775]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 21 04:48:52 crc kubenswrapper[4775]: --loglevel="${LOGLEVEL}" Mar 21 04:48:52 crc kubenswrapper[4775]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 21 04:48:52 crc kubenswrapper[4775]: > logger="UnhandledError" Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.845792 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 21 04:48:52 crc kubenswrapper[4775]: W0321 04:48:52.850251 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-14d8bf0b6d75a02942a6a502beb9579e7f6b9f1c6b1eb223e7d7c6b60d3a4fd4 WatchSource:0}: Error finding container 14d8bf0b6d75a02942a6a502beb9579e7f6b9f1c6b1eb223e7d7c6b60d3a4fd4: Status 404 returned error can't find the container with id 14d8bf0b6d75a02942a6a502beb9579e7f6b9f1c6b1eb223e7d7c6b60d3a4fd4 Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.852940 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 21 04:48:52 crc kubenswrapper[4775]: E0321 04:48:52.854111 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.875352 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.875399 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.875408 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.875439 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.875448 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:52Z","lastTransitionTime":"2026-03-21T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.977665 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.977712 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.977740 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.977761 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:52 crc kubenswrapper[4775]: I0321 04:48:52.977772 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:52Z","lastTransitionTime":"2026-03-21T04:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.080715 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.080765 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.080777 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.080795 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.080806 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:53Z","lastTransitionTime":"2026-03-21T04:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.182747 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.182793 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.182806 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.182824 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.182837 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:53Z","lastTransitionTime":"2026-03-21T04:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.218359 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.218519 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:48:54.218490122 +0000 UTC m=+87.194953746 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.218578 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.218614 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.218721 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.218789 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 04:48:54.21876851 +0000 UTC m=+87.195232174 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.218722 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.218823 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 04:48:54.218817331 +0000 UTC m=+87.195280955 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.285583 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.285632 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.285644 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.285660 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.285673 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:53Z","lastTransitionTime":"2026-03-21T04:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.319473 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.319516 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.319616 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.319636 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.319646 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.319694 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-21 04:48:54.319681409 +0000 UTC m=+87.296145033 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.319708 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.319744 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.319757 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.319821 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-21 04:48:54.319798252 +0000 UTC m=+87.296261916 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.387870 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.387915 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.387957 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.387974 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.387986 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:53Z","lastTransitionTime":"2026-03-21T04:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.490239 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.490297 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.490306 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.490321 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.490331 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:53Z","lastTransitionTime":"2026-03-21T04:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.576717 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"af3c6cc594d3df6f8481abbdfcef717c4a3b5771cfa7847295f4044b5982be48"} Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.577972 4775 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 21 04:48:53 crc kubenswrapper[4775]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 21 04:48:53 crc kubenswrapper[4775]: set -o allexport Mar 21 04:48:53 crc kubenswrapper[4775]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 21 04:48:53 crc kubenswrapper[4775]: source /etc/kubernetes/apiserver-url.env Mar 21 04:48:53 crc kubenswrapper[4775]: else Mar 21 04:48:53 crc kubenswrapper[4775]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 21 04:48:53 crc kubenswrapper[4775]: exit 1 Mar 21 04:48:53 crc kubenswrapper[4775]: fi Mar 21 04:48:53 crc kubenswrapper[4775]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 21 04:48:53 crc kubenswrapper[4775]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 21 04:48:53 crc kubenswrapper[4775]: > logger="UnhandledError" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.578919 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b1c7b006c84367f3a4790ee0ff97d47b510e4b6095d1cb4f1ca0271198041409"} Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.579209 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.580212 4775 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 21 04:48:53 crc kubenswrapper[4775]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 21 04:48:53 crc kubenswrapper[4775]: if [[ -f "/env/_master" ]]; then Mar 21 04:48:53 crc kubenswrapper[4775]: set -o allexport Mar 21 04:48:53 crc kubenswrapper[4775]: source "/env/_master" Mar 21 04:48:53 crc kubenswrapper[4775]: set +o allexport Mar 21 04:48:53 crc kubenswrapper[4775]: fi Mar 21 04:48:53 crc kubenswrapper[4775]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 21 04:48:53 crc kubenswrapper[4775]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 21 04:48:53 crc kubenswrapper[4775]: ho_enable="--enable-hybrid-overlay" Mar 21 04:48:53 crc kubenswrapper[4775]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 21 04:48:53 crc kubenswrapper[4775]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 21 04:48:53 crc kubenswrapper[4775]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 21 04:48:53 crc kubenswrapper[4775]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 21 04:48:53 crc kubenswrapper[4775]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 21 04:48:53 crc kubenswrapper[4775]: --webhook-host=127.0.0.1 \ Mar 21 04:48:53 crc kubenswrapper[4775]: --webhook-port=9743 \ Mar 21 04:48:53 crc kubenswrapper[4775]: ${ho_enable} \ Mar 21 04:48:53 crc kubenswrapper[4775]: --enable-interconnect \ Mar 21 04:48:53 crc kubenswrapper[4775]: --disable-approver \ Mar 21 04:48:53 crc kubenswrapper[4775]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 21 04:48:53 crc kubenswrapper[4775]: --wait-for-kubernetes-api=200s \ Mar 21 04:48:53 crc kubenswrapper[4775]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 21 04:48:53 crc kubenswrapper[4775]: --loglevel="${LOGLEVEL}" Mar 21 04:48:53 crc kubenswrapper[4775]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 21 04:48:53 crc kubenswrapper[4775]: > logger="UnhandledError" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.580430 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"14d8bf0b6d75a02942a6a502beb9579e7f6b9f1c6b1eb223e7d7c6b60d3a4fd4"} Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.581438 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.582585 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.582824 4775 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 21 04:48:53 crc kubenswrapper[4775]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 21 04:48:53 crc kubenswrapper[4775]: if [[ -f "/env/_master" ]]; then Mar 21 04:48:53 crc kubenswrapper[4775]: set -o allexport Mar 21 04:48:53 crc kubenswrapper[4775]: source "/env/_master" Mar 21 04:48:53 crc kubenswrapper[4775]: set +o allexport Mar 21 04:48:53 crc kubenswrapper[4775]: fi Mar 21 04:48:53 crc kubenswrapper[4775]: Mar 21 04:48:53 crc kubenswrapper[4775]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 21 04:48:53 crc kubenswrapper[4775]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 21 04:48:53 crc kubenswrapper[4775]: --disable-webhook \ Mar 21 04:48:53 crc kubenswrapper[4775]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 21 04:48:53 crc kubenswrapper[4775]: --loglevel="${LOGLEVEL}" Mar 21 04:48:53 crc kubenswrapper[4775]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 21 04:48:53 crc kubenswrapper[4775]: > logger="UnhandledError" Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.583986 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.587383 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.592580 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.592630 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.592645 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.592662 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.592673 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:53Z","lastTransitionTime":"2026-03-21T04:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.598339 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.609054 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.620316 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.632365 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.642970 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.655857 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.660530 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:48:53 crc kubenswrapper[4775]: E0321 04:48:53.660701 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.665416 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.666140 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.667710 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.670320 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.671818 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.672527 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.677383 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.678052 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.679473 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.680166 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.680568 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.680787 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.682088 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.682770 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.683886 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.684536 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.685752 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.686736 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.687411 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.688742 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.689663 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.690093 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.691526 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.692318 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.692833 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.695479 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.695948 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.697345 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.697398 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.697413 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.697431 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.697441 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:53Z","lastTransitionTime":"2026-03-21T04:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.697617 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.698537 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.699576 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.700306 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.701419 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.701933 4775 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.702051 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.702535 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.704511 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.705195 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.705645 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.707347 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.708465 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.709025 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.710543 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.711281 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.711825 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.712780 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.714046 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.715279 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.715998 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.717058 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.717722 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.718767 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.719634 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.720609 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.721096 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.721779 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.722213 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.722939 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.723544 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.724443 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.724929 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.799224 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.799285 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.799299 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.799317 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.799658 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:53Z","lastTransitionTime":"2026-03-21T04:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.901429 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.901506 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.901547 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.901562 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:53 crc kubenswrapper[4775]: I0321 04:48:53.901571 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:53Z","lastTransitionTime":"2026-03-21T04:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.003588 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.003642 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.003653 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.003670 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.003682 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:54Z","lastTransitionTime":"2026-03-21T04:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.109488 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.109540 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.109553 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.109573 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.109585 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:54Z","lastTransitionTime":"2026-03-21T04:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.212311 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.212385 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.212398 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.212412 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.212422 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:54Z","lastTransitionTime":"2026-03-21T04:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.228472 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.228553 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.228577 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:48:54 crc kubenswrapper[4775]: E0321 04:48:54.228678 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 04:48:54 crc kubenswrapper[4775]: E0321 04:48:54.228733 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 04:48:56.228717581 +0000 UTC m=+89.205181205 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 04:48:54 crc kubenswrapper[4775]: E0321 04:48:54.229102 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:48:56.229090502 +0000 UTC m=+89.205554136 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:48:54 crc kubenswrapper[4775]: E0321 04:48:54.229201 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 04:48:54 crc kubenswrapper[4775]: E0321 04:48:54.229238 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 04:48:56.229227515 +0000 UTC m=+89.205691139 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.315075 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.315150 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.315162 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.315180 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.315192 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:54Z","lastTransitionTime":"2026-03-21T04:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.329575 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.329620 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:48:54 crc kubenswrapper[4775]: E0321 04:48:54.329725 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 04:48:54 crc kubenswrapper[4775]: E0321 04:48:54.329739 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 04:48:54 crc kubenswrapper[4775]: E0321 04:48:54.329782 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:48:54 crc kubenswrapper[4775]: E0321 04:48:54.329818 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 04:48:54 crc kubenswrapper[4775]: E0321 04:48:54.329855 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 04:48:54 crc kubenswrapper[4775]: E0321 04:48:54.329869 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:48:54 crc kubenswrapper[4775]: E0321 04:48:54.329830 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-21 04:48:56.329817185 +0000 UTC m=+89.306280809 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:48:54 crc kubenswrapper[4775]: E0321 04:48:54.329941 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-21 04:48:56.329922488 +0000 UTC m=+89.306386182 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.417336 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.417385 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.417398 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.417415 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.417430 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:54Z","lastTransitionTime":"2026-03-21T04:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.417950 4775 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.519523 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.519564 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.519574 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.519609 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.519622 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:54Z","lastTransitionTime":"2026-03-21T04:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.622023 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.622183 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.622199 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.622222 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.622236 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:54Z","lastTransitionTime":"2026-03-21T04:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.661623 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.661764 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:48:54 crc kubenswrapper[4775]: E0321 04:48:54.661973 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:48:54 crc kubenswrapper[4775]: E0321 04:48:54.661783 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.724768 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.724809 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.724820 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.724835 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.724847 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:54Z","lastTransitionTime":"2026-03-21T04:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.828282 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.828318 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.828327 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.828340 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.828349 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:54Z","lastTransitionTime":"2026-03-21T04:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.931153 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.931196 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.931209 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.931226 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:54 crc kubenswrapper[4775]: I0321 04:48:54.931238 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:54Z","lastTransitionTime":"2026-03-21T04:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.033027 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.033078 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.033090 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.033142 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.033162 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:55Z","lastTransitionTime":"2026-03-21T04:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.135593 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.135630 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.135642 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.135658 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.135669 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:55Z","lastTransitionTime":"2026-03-21T04:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.238249 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.238299 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.238311 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.238329 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.238341 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:55Z","lastTransitionTime":"2026-03-21T04:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.340930 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.340961 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.340969 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.340982 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.340991 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:55Z","lastTransitionTime":"2026-03-21T04:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.442684 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.442713 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.442721 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.442733 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.442743 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:55Z","lastTransitionTime":"2026-03-21T04:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.544925 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.545018 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.545031 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.545073 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.545083 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:55Z","lastTransitionTime":"2026-03-21T04:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.647784 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.647848 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.647863 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.647901 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.647915 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:55Z","lastTransitionTime":"2026-03-21T04:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.661480 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:48:55 crc kubenswrapper[4775]: E0321 04:48:55.661716 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.750443 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.750499 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.750512 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.750528 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.750539 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:55Z","lastTransitionTime":"2026-03-21T04:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.853267 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.853302 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.853311 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.853324 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.853332 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:55Z","lastTransitionTime":"2026-03-21T04:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.955827 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.955862 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.955872 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.955885 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:55 crc kubenswrapper[4775]: I0321 04:48:55.955894 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:55Z","lastTransitionTime":"2026-03-21T04:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.057634 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.057674 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.057685 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.057700 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.057710 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:56Z","lastTransitionTime":"2026-03-21T04:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.159595 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.159636 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.159645 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.159658 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.159667 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:56Z","lastTransitionTime":"2026-03-21T04:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.244779 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.244861 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.244885 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:48:56 crc kubenswrapper[4775]: E0321 04:48:56.244963 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 04:48:56 crc kubenswrapper[4775]: E0321 04:48:56.245008 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 04:49:00.24499518 +0000 UTC m=+93.221458804 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 04:48:56 crc kubenswrapper[4775]: E0321 04:48:56.245411 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 04:48:56 crc kubenswrapper[4775]: E0321 04:48:56.245506 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:49:00.245472393 +0000 UTC m=+93.221936017 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:48:56 crc kubenswrapper[4775]: E0321 04:48:56.245535 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 04:49:00.245528615 +0000 UTC m=+93.221992239 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.262080 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.262311 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.262385 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.262464 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.262532 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:56Z","lastTransitionTime":"2026-03-21T04:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.346109 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.346186 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:48:56 crc kubenswrapper[4775]: E0321 04:48:56.346314 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 04:48:56 crc kubenswrapper[4775]: E0321 04:48:56.346336 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 04:48:56 crc kubenswrapper[4775]: E0321 04:48:56.346347 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:48:56 crc kubenswrapper[4775]: E0321 04:48:56.346402 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-21 04:49:00.346384882 +0000 UTC m=+93.322848506 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:48:56 crc kubenswrapper[4775]: E0321 04:48:56.346314 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 04:48:56 crc kubenswrapper[4775]: E0321 04:48:56.346767 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 04:48:56 crc kubenswrapper[4775]: E0321 04:48:56.346839 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:48:56 crc kubenswrapper[4775]: E0321 04:48:56.346930 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-21 04:49:00.346919267 +0000 UTC m=+93.323382891 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.364882 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.364917 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.364929 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.364946 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.364967 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:56Z","lastTransitionTime":"2026-03-21T04:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.467162 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.467215 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.467229 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.467245 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.467256 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:56Z","lastTransitionTime":"2026-03-21T04:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.569064 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.569102 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.569139 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.569155 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.569165 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:56Z","lastTransitionTime":"2026-03-21T04:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.660823 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.660867 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:48:56 crc kubenswrapper[4775]: E0321 04:48:56.660977 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:48:56 crc kubenswrapper[4775]: E0321 04:48:56.661362 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.671495 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.671522 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.671530 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.671544 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.671553 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:56Z","lastTransitionTime":"2026-03-21T04:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.773619 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.773670 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.773683 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.773708 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.773725 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:56Z","lastTransitionTime":"2026-03-21T04:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.876050 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.876086 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.876095 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.876109 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.876131 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:56Z","lastTransitionTime":"2026-03-21T04:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.981311 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.981361 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.981479 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.982053 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:56 crc kubenswrapper[4775]: I0321 04:48:56.982138 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:56Z","lastTransitionTime":"2026-03-21T04:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.084691 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.084729 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.084737 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.084750 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.084759 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:57Z","lastTransitionTime":"2026-03-21T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.187534 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.187573 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.187582 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.187595 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.187605 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:57Z","lastTransitionTime":"2026-03-21T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.289792 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.289868 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.289882 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.289898 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.289908 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:57Z","lastTransitionTime":"2026-03-21T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.392340 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.392955 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.392970 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.392986 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.392996 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:57Z","lastTransitionTime":"2026-03-21T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.494740 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.494783 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.494792 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.494807 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.494816 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:57Z","lastTransitionTime":"2026-03-21T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.596361 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.596400 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.596411 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.596430 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.596442 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:57Z","lastTransitionTime":"2026-03-21T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.660515 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:48:57 crc kubenswrapper[4775]: E0321 04:48:57.660667 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.683349 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.692911 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.698474 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.698641 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.698744 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.698830 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.698929 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:57Z","lastTransitionTime":"2026-03-21T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.703090 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.712645 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.721437 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.731583 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.742358 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.801149 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.801199 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.801214 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.801236 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.801251 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:57Z","lastTransitionTime":"2026-03-21T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.882454 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.882760 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.882845 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.882935 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.883022 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:57Z","lastTransitionTime":"2026-03-21T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:57 crc kubenswrapper[4775]: E0321 04:48:57.892829 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.896695 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.896838 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.896930 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.897031 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.897151 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:57Z","lastTransitionTime":"2026-03-21T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:57 crc kubenswrapper[4775]: E0321 04:48:57.907399 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.911359 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.911404 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.911413 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.911428 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.911437 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:57Z","lastTransitionTime":"2026-03-21T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:57 crc kubenswrapper[4775]: E0321 04:48:57.921532 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.926044 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.926094 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.926104 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.926164 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.926176 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:57Z","lastTransitionTime":"2026-03-21T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:57 crc kubenswrapper[4775]: E0321 04:48:57.936495 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.939700 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.939822 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.939920 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.940044 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.940182 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:57Z","lastTransitionTime":"2026-03-21T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:57 crc kubenswrapper[4775]: E0321 04:48:57.950910 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:48:57 crc kubenswrapper[4775]: E0321 04:48:57.951349 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.952875 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.952907 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.952919 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.952934 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:57 crc kubenswrapper[4775]: I0321 04:48:57.952947 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:57Z","lastTransitionTime":"2026-03-21T04:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.055023 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.055276 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.055388 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.055484 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.055571 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:58Z","lastTransitionTime":"2026-03-21T04:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.158227 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.158517 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.158589 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.158655 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.158713 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:58Z","lastTransitionTime":"2026-03-21T04:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.261944 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.261984 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.261996 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.262010 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.262021 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:58Z","lastTransitionTime":"2026-03-21T04:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.365239 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.365282 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.365292 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.365309 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.365320 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:58Z","lastTransitionTime":"2026-03-21T04:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.467824 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.468138 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.468215 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.468316 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.468401 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:58Z","lastTransitionTime":"2026-03-21T04:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.570795 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.571057 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.571194 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.571321 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.571413 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:58Z","lastTransitionTime":"2026-03-21T04:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.660938 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:48:58 crc kubenswrapper[4775]: E0321 04:48:58.661318 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.660956 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:48:58 crc kubenswrapper[4775]: E0321 04:48:58.661666 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.670472 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.670941 4775 scope.go:117] "RemoveContainer" containerID="85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4" Mar 21 04:48:58 crc kubenswrapper[4775]: E0321 04:48:58.671300 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.674012 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.674042 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.674053 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.674076 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.674088 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:58Z","lastTransitionTime":"2026-03-21T04:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.776102 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.776206 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.776222 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.776237 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.776273 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:58Z","lastTransitionTime":"2026-03-21T04:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.879634 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.879888 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.879958 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.880040 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.880151 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:58Z","lastTransitionTime":"2026-03-21T04:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.982535 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.982843 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.983023 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.983234 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:58 crc kubenswrapper[4775]: I0321 04:48:58.983404 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:58Z","lastTransitionTime":"2026-03-21T04:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.085605 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.085676 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.085689 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.085705 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.085716 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:59Z","lastTransitionTime":"2026-03-21T04:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.188177 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.188219 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.188232 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.188249 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.188261 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:59Z","lastTransitionTime":"2026-03-21T04:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.290294 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.290336 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.290346 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.290384 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.290400 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:59Z","lastTransitionTime":"2026-03-21T04:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.392147 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.392261 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.392279 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.392300 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.392311 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:59Z","lastTransitionTime":"2026-03-21T04:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.494312 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.494353 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.494364 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.494379 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.494389 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:59Z","lastTransitionTime":"2026-03-21T04:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.594821 4775 scope.go:117] "RemoveContainer" containerID="85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4" Mar 21 04:48:59 crc kubenswrapper[4775]: E0321 04:48:59.595008 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.596423 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.596458 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.596468 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.596482 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.596491 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:59Z","lastTransitionTime":"2026-03-21T04:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.660731 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:48:59 crc kubenswrapper[4775]: E0321 04:48:59.660845 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.699108 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.699167 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.699175 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.699189 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.699198 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:59Z","lastTransitionTime":"2026-03-21T04:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.801240 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.801273 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.801283 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.801295 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.801304 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:59Z","lastTransitionTime":"2026-03-21T04:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.905637 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.905691 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.905701 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.905716 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:48:59 crc kubenswrapper[4775]: I0321 04:48:59.905728 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:48:59Z","lastTransitionTime":"2026-03-21T04:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.008434 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.008485 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.008498 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.008516 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.008529 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:00Z","lastTransitionTime":"2026-03-21T04:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.110968 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.111106 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.111154 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.111173 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.111185 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:00Z","lastTransitionTime":"2026-03-21T04:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.214179 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.214214 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.214224 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.214240 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.214252 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:00Z","lastTransitionTime":"2026-03-21T04:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.278619 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.278703 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.278734 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:00 crc kubenswrapper[4775]: E0321 04:49:00.278849 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 04:49:00 crc kubenswrapper[4775]: E0321 04:49:00.278870 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:49:08.278838594 +0000 UTC m=+101.255302218 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:49:00 crc kubenswrapper[4775]: E0321 04:49:00.278907 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 04:49:08.278897886 +0000 UTC m=+101.255361620 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 04:49:00 crc kubenswrapper[4775]: E0321 04:49:00.278988 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 04:49:00 crc kubenswrapper[4775]: E0321 04:49:00.279071 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 04:49:08.27904949 +0000 UTC m=+101.255513194 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.316896 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.316925 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.316933 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.316945 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.316954 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:00Z","lastTransitionTime":"2026-03-21T04:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.379930 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.379976 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:00 crc kubenswrapper[4775]: E0321 04:49:00.380093 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 04:49:00 crc kubenswrapper[4775]: E0321 04:49:00.380139 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 04:49:00 crc kubenswrapper[4775]: E0321 04:49:00.380152 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:49:00 crc kubenswrapper[4775]: E0321 04:49:00.380195 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-21 04:49:08.380180415 +0000 UTC m=+101.356644029 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:49:00 crc kubenswrapper[4775]: E0321 04:49:00.380512 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 04:49:00 crc kubenswrapper[4775]: E0321 04:49:00.380532 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 04:49:00 crc kubenswrapper[4775]: E0321 04:49:00.380543 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:49:00 crc kubenswrapper[4775]: E0321 04:49:00.380570 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-21 04:49:08.380563396 +0000 UTC m=+101.357027020 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.425344 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.425407 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.425420 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.425440 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.425455 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:00Z","lastTransitionTime":"2026-03-21T04:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.528358 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.528392 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.528401 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.528414 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.528424 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:00Z","lastTransitionTime":"2026-03-21T04:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.631677 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.631723 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.631734 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.631753 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.631764 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:00Z","lastTransitionTime":"2026-03-21T04:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.661282 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.661361 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:00 crc kubenswrapper[4775]: E0321 04:49:00.661402 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:00 crc kubenswrapper[4775]: E0321 04:49:00.661494 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.734348 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.734386 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.734397 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.734410 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.734421 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:00Z","lastTransitionTime":"2026-03-21T04:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.836590 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.836661 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.836675 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.836694 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.836708 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:00Z","lastTransitionTime":"2026-03-21T04:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.938448 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.938492 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.938502 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.938517 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:00 crc kubenswrapper[4775]: I0321 04:49:00.938526 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:00Z","lastTransitionTime":"2026-03-21T04:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.041938 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.041978 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.041987 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.042001 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.042009 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:01Z","lastTransitionTime":"2026-03-21T04:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.145653 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.145704 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.145719 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.145736 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.145749 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:01Z","lastTransitionTime":"2026-03-21T04:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.247391 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.247421 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.247431 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.247443 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.247452 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:01Z","lastTransitionTime":"2026-03-21T04:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.349752 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.349818 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.349830 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.349847 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.349858 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:01Z","lastTransitionTime":"2026-03-21T04:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.452659 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.452706 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.452718 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.452736 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.452754 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:01Z","lastTransitionTime":"2026-03-21T04:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.555172 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.555221 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.555237 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.555256 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.555268 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:01Z","lastTransitionTime":"2026-03-21T04:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.657872 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.657917 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.657928 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.657944 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.657955 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:01Z","lastTransitionTime":"2026-03-21T04:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.661302 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:01 crc kubenswrapper[4775]: E0321 04:49:01.661462 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.759619 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.759660 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.759678 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.759697 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.759709 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:01Z","lastTransitionTime":"2026-03-21T04:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.834066 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-khh7x"] Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.834406 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-khh7x" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.837191 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.837196 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.838596 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.845043 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.856672 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.861284 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.861322 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.861334 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.861350 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.861362 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:01Z","lastTransitionTime":"2026-03-21T04:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.867356 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.877665 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.887087 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.896715 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.901136 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwtgh\" (UniqueName: \"kubernetes.io/projected/70cca5ae-9de4-4933-a6f8-4a23ab711bbf-kube-api-access-bwtgh\") pod \"node-resolver-khh7x\" (UID: \"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\") " pod="openshift-dns/node-resolver-khh7x" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.901180 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/70cca5ae-9de4-4933-a6f8-4a23ab711bbf-hosts-file\") pod \"node-resolver-khh7x\" (UID: \"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\") " pod="openshift-dns/node-resolver-khh7x" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.911218 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.921109 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.926647 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.964915 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.964964 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.964980 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.964997 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:01 crc kubenswrapper[4775]: I0321 04:49:01.965010 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:01Z","lastTransitionTime":"2026-03-21T04:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.001953 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/70cca5ae-9de4-4933-a6f8-4a23ab711bbf-hosts-file\") pod \"node-resolver-khh7x\" (UID: \"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\") " pod="openshift-dns/node-resolver-khh7x" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.002029 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwtgh\" (UniqueName: \"kubernetes.io/projected/70cca5ae-9de4-4933-a6f8-4a23ab711bbf-kube-api-access-bwtgh\") pod \"node-resolver-khh7x\" (UID: \"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\") " pod="openshift-dns/node-resolver-khh7x" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.002179 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/70cca5ae-9de4-4933-a6f8-4a23ab711bbf-hosts-file\") pod \"node-resolver-khh7x\" (UID: \"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\") " pod="openshift-dns/node-resolver-khh7x" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.020576 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwtgh\" (UniqueName: \"kubernetes.io/projected/70cca5ae-9de4-4933-a6f8-4a23ab711bbf-kube-api-access-bwtgh\") pod \"node-resolver-khh7x\" (UID: \"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\") " pod="openshift-dns/node-resolver-khh7x" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.067278 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.067316 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.067325 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.067338 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.067346 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:02Z","lastTransitionTime":"2026-03-21T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.149246 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-qc7hn"] Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.150621 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-khh7x" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.150782 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.151384 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-556rg"] Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.151639 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-kldzh"] Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.151918 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.152141 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.155257 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.155528 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.155854 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.155988 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.156335 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.156584 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.156640 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.156861 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.157013 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.157484 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.157672 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.157796 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.170975 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.172446 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.172477 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.172490 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.172505 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.172516 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:02Z","lastTransitionTime":"2026-03-21T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.178375 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.192042 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.199887 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.204720 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/957cde70-ca20-438a-a4bf-42481dddb2db-tuning-conf-dir\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.204771 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-multus-conf-dir\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.204793 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cffcf487-ef41-4395-81eb-e5e6358f4a32-proxy-tls\") pod \"machine-config-daemon-qc7hn\" (UID: \"cffcf487-ef41-4395-81eb-e5e6358f4a32\") " pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.204811 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/957cde70-ca20-438a-a4bf-42481dddb2db-system-cni-dir\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.204833 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e77ec218-42da-4f07-b214-184c4f3b20f3-cni-binary-copy\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.204870 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/cffcf487-ef41-4395-81eb-e5e6358f4a32-rootfs\") pod \"machine-config-daemon-qc7hn\" (UID: \"cffcf487-ef41-4395-81eb-e5e6358f4a32\") " pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.204889 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-multus-socket-dir-parent\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.204909 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vft22\" (UniqueName: \"kubernetes.io/projected/e77ec218-42da-4f07-b214-184c4f3b20f3-kube-api-access-vft22\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.204929 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-etc-kubernetes\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.204965 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/957cde70-ca20-438a-a4bf-42481dddb2db-cni-binary-copy\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.204985 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-hostroot\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.205004 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pxr5\" (UniqueName: \"kubernetes.io/projected/cffcf487-ef41-4395-81eb-e5e6358f4a32-kube-api-access-9pxr5\") pod \"machine-config-daemon-qc7hn\" (UID: \"cffcf487-ef41-4395-81eb-e5e6358f4a32\") " pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.205024 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-host-run-multus-certs\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.205043 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-multus-cni-dir\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.205062 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-os-release\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.205080 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-host-run-netns\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.205100 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-host-var-lib-cni-multus\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.205141 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-host-var-lib-kubelet\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.205163 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-host-run-k8s-cni-cncf-io\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.205183 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-host-var-lib-cni-bin\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.205202 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-cnibin\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.205224 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e77ec218-42da-4f07-b214-184c4f3b20f3-multus-daemon-config\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.205247 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqnl6\" (UniqueName: \"kubernetes.io/projected/957cde70-ca20-438a-a4bf-42481dddb2db-kube-api-access-sqnl6\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.205266 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-system-cni-dir\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.205287 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cffcf487-ef41-4395-81eb-e5e6358f4a32-mcd-auth-proxy-config\") pod \"machine-config-daemon-qc7hn\" (UID: \"cffcf487-ef41-4395-81eb-e5e6358f4a32\") " pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.205309 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/957cde70-ca20-438a-a4bf-42481dddb2db-cnibin\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.205330 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/957cde70-ca20-438a-a4bf-42481dddb2db-os-release\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.205350 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/957cde70-ca20-438a-a4bf-42481dddb2db-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.208996 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.217595 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.227568 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.235150 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.246814 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.259688 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.274781 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.274816 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.274827 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.274843 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.274852 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:02Z","lastTransitionTime":"2026-03-21T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.284446 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.294551 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.305827 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.306428 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/cffcf487-ef41-4395-81eb-e5e6358f4a32-rootfs\") pod \"machine-config-daemon-qc7hn\" (UID: \"cffcf487-ef41-4395-81eb-e5e6358f4a32\") " pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.306521 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-multus-socket-dir-parent\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.307236 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vft22\" (UniqueName: \"kubernetes.io/projected/e77ec218-42da-4f07-b214-184c4f3b20f3-kube-api-access-vft22\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.306521 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/cffcf487-ef41-4395-81eb-e5e6358f4a32-rootfs\") pod \"machine-config-daemon-qc7hn\" (UID: \"cffcf487-ef41-4395-81eb-e5e6358f4a32\") " pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.306669 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-multus-socket-dir-parent\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.307517 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-etc-kubernetes\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.307559 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/957cde70-ca20-438a-a4bf-42481dddb2db-cni-binary-copy\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.307606 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-hostroot\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.307633 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-etc-kubernetes\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.307921 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-hostroot\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.308369 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pxr5\" (UniqueName: \"kubernetes.io/projected/cffcf487-ef41-4395-81eb-e5e6358f4a32-kube-api-access-9pxr5\") pod \"machine-config-daemon-qc7hn\" (UID: \"cffcf487-ef41-4395-81eb-e5e6358f4a32\") " pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.308523 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-host-run-multus-certs\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309343 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-host-run-multus-certs\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309384 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-multus-cni-dir\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309410 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-os-release\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309434 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-host-run-netns\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309455 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-host-var-lib-cni-multus\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309494 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-multus-cni-dir\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309516 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-host-var-lib-kubelet\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309539 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-host-run-k8s-cni-cncf-io\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309559 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-host-var-lib-cni-bin\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309581 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-cnibin\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309603 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e77ec218-42da-4f07-b214-184c4f3b20f3-multus-daemon-config\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309605 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-host-var-lib-kubelet\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309628 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqnl6\" (UniqueName: \"kubernetes.io/projected/957cde70-ca20-438a-a4bf-42481dddb2db-kube-api-access-sqnl6\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309629 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-host-run-k8s-cni-cncf-io\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309682 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-system-cni-dir\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309609 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-host-run-netns\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309750 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cffcf487-ef41-4395-81eb-e5e6358f4a32-mcd-auth-proxy-config\") pod \"machine-config-daemon-qc7hn\" (UID: \"cffcf487-ef41-4395-81eb-e5e6358f4a32\") " pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309758 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-cnibin\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309781 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/957cde70-ca20-438a-a4bf-42481dddb2db-cnibin\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309813 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/957cde70-ca20-438a-a4bf-42481dddb2db-cnibin\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309929 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/957cde70-ca20-438a-a4bf-42481dddb2db-os-release\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.310073 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/957cde70-ca20-438a-a4bf-42481dddb2db-os-release\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309695 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-os-release\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309866 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-system-cni-dir\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.309989 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/957cde70-ca20-438a-a4bf-42481dddb2db-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.310289 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/957cde70-ca20-438a-a4bf-42481dddb2db-tuning-conf-dir\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.310312 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-multus-conf-dir\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.310352 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cffcf487-ef41-4395-81eb-e5e6358f4a32-proxy-tls\") pod \"machine-config-daemon-qc7hn\" (UID: \"cffcf487-ef41-4395-81eb-e5e6358f4a32\") " pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.310374 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/957cde70-ca20-438a-a4bf-42481dddb2db-system-cni-dir\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.310396 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e77ec218-42da-4f07-b214-184c4f3b20f3-cni-binary-copy\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.310575 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cffcf487-ef41-4395-81eb-e5e6358f4a32-mcd-auth-proxy-config\") pod \"machine-config-daemon-qc7hn\" (UID: \"cffcf487-ef41-4395-81eb-e5e6358f4a32\") " pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.310632 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e77ec218-42da-4f07-b214-184c4f3b20f3-multus-daemon-config\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.310689 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-multus-conf-dir\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.310983 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/957cde70-ca20-438a-a4bf-42481dddb2db-system-cni-dir\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.311010 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e77ec218-42da-4f07-b214-184c4f3b20f3-cni-binary-copy\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.311301 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-host-var-lib-cni-bin\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.311345 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e77ec218-42da-4f07-b214-184c4f3b20f3-host-var-lib-cni-multus\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.312423 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/957cde70-ca20-438a-a4bf-42481dddb2db-tuning-conf-dir\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.318320 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/957cde70-ca20-438a-a4bf-42481dddb2db-cni-binary-copy\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.318726 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cffcf487-ef41-4395-81eb-e5e6358f4a32-proxy-tls\") pod \"machine-config-daemon-qc7hn\" (UID: \"cffcf487-ef41-4395-81eb-e5e6358f4a32\") " pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.322291 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/957cde70-ca20-438a-a4bf-42481dddb2db-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.322389 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.325240 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vft22\" (UniqueName: \"kubernetes.io/projected/e77ec218-42da-4f07-b214-184c4f3b20f3-kube-api-access-vft22\") pod \"multus-556rg\" (UID: \"e77ec218-42da-4f07-b214-184c4f3b20f3\") " pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.325427 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pxr5\" (UniqueName: \"kubernetes.io/projected/cffcf487-ef41-4395-81eb-e5e6358f4a32-kube-api-access-9pxr5\") pod \"machine-config-daemon-qc7hn\" (UID: \"cffcf487-ef41-4395-81eb-e5e6358f4a32\") " pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.328981 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqnl6\" (UniqueName: \"kubernetes.io/projected/957cde70-ca20-438a-a4bf-42481dddb2db-kube-api-access-sqnl6\") pod \"multus-additional-cni-plugins-kldzh\" (UID: \"957cde70-ca20-438a-a4bf-42481dddb2db\") " pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.334345 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.345617 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.353836 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.383548 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.383592 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.383604 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.383620 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.383630 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:02Z","lastTransitionTime":"2026-03-21T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.389187 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.400658 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.412190 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.423753 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.433251 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.480907 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.488088 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-556rg" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.495732 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-kldzh" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.497916 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.497937 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.497946 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.497960 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.497971 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:02Z","lastTransitionTime":"2026-03-21T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.520440 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mzqtk"] Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.521225 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: W0321 04:49:02.523433 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod957cde70_ca20_438a_a4bf_42481dddb2db.slice/crio-301da03471c41946481eae1a4d5aba941d0f6e54e94da5ce2e7db5cc0688799d WatchSource:0}: Error finding container 301da03471c41946481eae1a4d5aba941d0f6e54e94da5ce2e7db5cc0688799d: Status 404 returned error can't find the container with id 301da03471c41946481eae1a4d5aba941d0f6e54e94da5ce2e7db5cc0688799d Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.524035 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.524186 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.524330 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.524557 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.525107 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.525242 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.532017 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.537452 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.547185 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.566805 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.577952 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.586648 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.595977 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.603555 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.605155 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.605182 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.605297 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.605310 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:02Z","lastTransitionTime":"2026-03-21T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.605516 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" event={"ID":"957cde70-ca20-438a-a4bf-42481dddb2db","Type":"ContainerStarted","Data":"301da03471c41946481eae1a4d5aba941d0f6e54e94da5ce2e7db5cc0688799d"} Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.606501 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.607841 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerStarted","Data":"7ef2a3b6220db4bc8aca03b1b3bd662eba9eefcf7c61a83b758d2c0438f4c19b"} Mar 21 04:49:02 crc kubenswrapper[4775]: W0321 04:49:02.616546 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode77ec218_42da_4f07_b214_184c4f3b20f3.slice/crio-a4b14fb323638d862a85c09bb4e48af7322efaab141d2c6f6e3a94ca0a74466d WatchSource:0}: Error finding container a4b14fb323638d862a85c09bb4e48af7322efaab141d2c6f6e3a94ca0a74466d: Status 404 returned error can't find the container with id a4b14fb323638d862a85c09bb4e48af7322efaab141d2c6f6e3a94ca0a74466d Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.616605 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-khh7x" event={"ID":"70cca5ae-9de4-4933-a6f8-4a23ab711bbf","Type":"ContainerStarted","Data":"6e58e7b588aafb219cf072c2bd4bc4bbc5b8413034949e8002754ff3bb657735"} Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.616873 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-var-lib-openvswitch\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.616899 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-etc-openvswitch\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.616919 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-run-openvswitch\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.616934 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-node-log\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.616947 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-cni-netd\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.616961 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-log-socket\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.616974 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-cni-bin\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.616990 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.617015 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-kubelet\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.617030 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-run-ovn\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.617056 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a69d31f5-deeb-4860-be96-ed5547831685-ovn-node-metrics-cert\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.617101 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-run-ovn-kubernetes\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.617143 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-systemd-units\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.617162 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a69d31f5-deeb-4860-be96-ed5547831685-ovnkube-config\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.617271 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a69d31f5-deeb-4860-be96-ed5547831685-ovnkube-script-lib\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.617299 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7w6b\" (UniqueName: \"kubernetes.io/projected/a69d31f5-deeb-4860-be96-ed5547831685-kube-api-access-h7w6b\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.617330 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-run-netns\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.617345 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a69d31f5-deeb-4860-be96-ed5547831685-env-overrides\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.617363 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-run-systemd\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.617377 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-slash\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.617434 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.629216 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.646836 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.657508 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.660358 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:02 crc kubenswrapper[4775]: E0321 04:49:02.660443 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.660708 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:02 crc kubenswrapper[4775]: E0321 04:49:02.660759 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.666501 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.675452 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.710735 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.710776 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.710788 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.710805 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.710818 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:02Z","lastTransitionTime":"2026-03-21T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.718750 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7w6b\" (UniqueName: \"kubernetes.io/projected/a69d31f5-deeb-4860-be96-ed5547831685-kube-api-access-h7w6b\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.718786 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a69d31f5-deeb-4860-be96-ed5547831685-ovnkube-config\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.718803 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a69d31f5-deeb-4860-be96-ed5547831685-ovnkube-script-lib\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.718818 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a69d31f5-deeb-4860-be96-ed5547831685-env-overrides\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.718863 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-run-netns\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.718880 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-run-systemd\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.718895 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-slash\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.718913 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-var-lib-openvswitch\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.718933 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-etc-openvswitch\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.718949 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-run-openvswitch\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.718964 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-node-log\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.718979 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-cni-netd\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.719000 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-log-socket\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.719014 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-cni-bin\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.719032 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.719056 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-kubelet\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.719072 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-run-ovn\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.719087 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a69d31f5-deeb-4860-be96-ed5547831685-ovn-node-metrics-cert\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.719110 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-run-ovn-kubernetes\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.719148 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-systemd-units\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.719210 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-systemd-units\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.720212 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a69d31f5-deeb-4860-be96-ed5547831685-ovnkube-config\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.720395 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-node-log\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.720438 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-var-lib-openvswitch\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.720469 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-etc-openvswitch\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.720531 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-run-openvswitch\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.720562 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-run-netns\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.720664 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-slash\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.720748 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.720789 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-cni-netd\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.720812 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-cni-bin\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.720845 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-kubelet\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.720884 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-run-ovn-kubernetes\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.720769 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-log-socket\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.720979 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-run-ovn\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.721005 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a69d31f5-deeb-4860-be96-ed5547831685-env-overrides\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.721433 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-run-systemd\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.721658 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a69d31f5-deeb-4860-be96-ed5547831685-ovnkube-script-lib\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.723875 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a69d31f5-deeb-4860-be96-ed5547831685-ovn-node-metrics-cert\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.737743 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7w6b\" (UniqueName: \"kubernetes.io/projected/a69d31f5-deeb-4860-be96-ed5547831685-kube-api-access-h7w6b\") pod \"ovnkube-node-mzqtk\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.813715 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.813757 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.813768 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.813782 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.813792 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:02Z","lastTransitionTime":"2026-03-21T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.833344 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:02 crc kubenswrapper[4775]: W0321 04:49:02.851664 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda69d31f5_deeb_4860_be96_ed5547831685.slice/crio-a053283b74d1089f3887c9fbc93ee3063db45e363837d9c9231a5cf45771e07b WatchSource:0}: Error finding container a053283b74d1089f3887c9fbc93ee3063db45e363837d9c9231a5cf45771e07b: Status 404 returned error can't find the container with id a053283b74d1089f3887c9fbc93ee3063db45e363837d9c9231a5cf45771e07b Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.916735 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.916772 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.916784 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.916800 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:02 crc kubenswrapper[4775]: I0321 04:49:02.916812 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:02Z","lastTransitionTime":"2026-03-21T04:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.019604 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.019661 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.019672 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.019691 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.019703 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:03Z","lastTransitionTime":"2026-03-21T04:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.122133 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.122185 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.122198 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.122214 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.122225 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:03Z","lastTransitionTime":"2026-03-21T04:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.225289 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.225337 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.225348 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.225395 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.225410 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:03Z","lastTransitionTime":"2026-03-21T04:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.327441 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.327672 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.327799 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.328076 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.328238 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:03Z","lastTransitionTime":"2026-03-21T04:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.431081 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.431362 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.431490 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.431621 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.431689 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:03Z","lastTransitionTime":"2026-03-21T04:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.534346 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.534709 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.534817 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.534906 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.534984 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:03Z","lastTransitionTime":"2026-03-21T04:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.620768 4775 generic.go:334] "Generic (PLEG): container finished" podID="a69d31f5-deeb-4860-be96-ed5547831685" containerID="effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6" exitCode=0 Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.620839 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerDied","Data":"effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6"} Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.620866 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerStarted","Data":"a053283b74d1089f3887c9fbc93ee3063db45e363837d9c9231a5cf45771e07b"} Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.622828 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-556rg" event={"ID":"e77ec218-42da-4f07-b214-184c4f3b20f3","Type":"ContainerStarted","Data":"77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf"} Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.622913 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-556rg" event={"ID":"e77ec218-42da-4f07-b214-184c4f3b20f3","Type":"ContainerStarted","Data":"a4b14fb323638d862a85c09bb4e48af7322efaab141d2c6f6e3a94ca0a74466d"} Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.624217 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" event={"ID":"957cde70-ca20-438a-a4bf-42481dddb2db","Type":"ContainerStarted","Data":"10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18"} Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.626478 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerStarted","Data":"3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c"} Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.626601 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerStarted","Data":"94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee"} Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.628579 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-khh7x" event={"ID":"70cca5ae-9de4-4933-a6f8-4a23ab711bbf","Type":"ContainerStarted","Data":"cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e"} Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.637795 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.637840 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.637850 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.637864 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.637874 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:03Z","lastTransitionTime":"2026-03-21T04:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.646414 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.656547 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.660329 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:03 crc kubenswrapper[4775]: E0321 04:49:03.660450 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.677416 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.687660 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.697367 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.705277 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.714212 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.722351 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.733814 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.740623 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.740665 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.740674 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.740692 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.740702 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:03Z","lastTransitionTime":"2026-03-21T04:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.743443 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.752552 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.760602 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.767798 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.774526 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.784747 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.797094 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.808017 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.818654 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.828443 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.844663 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.844699 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.844709 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.844722 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.844731 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:03Z","lastTransitionTime":"2026-03-21T04:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.850502 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.866938 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.888883 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.904162 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.915881 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.926184 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.938498 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.947171 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.947199 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.947207 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.947220 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:03 crc kubenswrapper[4775]: I0321 04:49:03.947228 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:03Z","lastTransitionTime":"2026-03-21T04:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.049852 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.049891 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.049900 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.049915 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.049928 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:04Z","lastTransitionTime":"2026-03-21T04:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.153719 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.153768 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.153778 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.153794 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.153806 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:04Z","lastTransitionTime":"2026-03-21T04:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.256143 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.256191 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.256202 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.256219 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.256232 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:04Z","lastTransitionTime":"2026-03-21T04:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.359620 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.359677 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.359686 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.359700 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.359710 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:04Z","lastTransitionTime":"2026-03-21T04:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.462248 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.462279 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.462289 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.462302 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.462310 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:04Z","lastTransitionTime":"2026-03-21T04:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.564652 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.564682 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.564691 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.564703 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.564714 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:04Z","lastTransitionTime":"2026-03-21T04:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.632368 4775 generic.go:334] "Generic (PLEG): container finished" podID="957cde70-ca20-438a-a4bf-42481dddb2db" containerID="10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18" exitCode=0 Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.632987 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" event={"ID":"957cde70-ca20-438a-a4bf-42481dddb2db","Type":"ContainerDied","Data":"10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18"} Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.638085 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerStarted","Data":"6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc"} Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.638153 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerStarted","Data":"6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c"} Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.638169 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerStarted","Data":"5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde"} Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.649951 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.660916 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.660926 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:04 crc kubenswrapper[4775]: E0321 04:49:04.661011 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:04 crc kubenswrapper[4775]: E0321 04:49:04.661072 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.667269 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.667781 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.667809 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.667820 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.667836 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.667848 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:04Z","lastTransitionTime":"2026-03-21T04:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.675450 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.683054 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-kr988"] Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.683435 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-kr988" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.685585 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.685837 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.686080 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.686180 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.686239 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.693662 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.705061 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.714169 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.723659 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.732888 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.741637 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.756827 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.766682 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.770006 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.770031 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.770039 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.770053 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.770061 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:04Z","lastTransitionTime":"2026-03-21T04:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.778192 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.786154 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cc87971f-e8fc-454d-8513-957a0bbad389-serviceca\") pod \"node-ca-kr988\" (UID: \"cc87971f-e8fc-454d-8513-957a0bbad389\") " pod="openshift-image-registry/node-ca-kr988" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.786220 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2kj8\" (UniqueName: \"kubernetes.io/projected/cc87971f-e8fc-454d-8513-957a0bbad389-kube-api-access-q2kj8\") pod \"node-ca-kr988\" (UID: \"cc87971f-e8fc-454d-8513-957a0bbad389\") " pod="openshift-image-registry/node-ca-kr988" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.786246 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc87971f-e8fc-454d-8513-957a0bbad389-host\") pod \"node-ca-kr988\" (UID: \"cc87971f-e8fc-454d-8513-957a0bbad389\") " pod="openshift-image-registry/node-ca-kr988" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.793788 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.801332 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.818179 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.851760 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.866749 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.871729 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.871805 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.871819 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.871835 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.871845 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:04Z","lastTransitionTime":"2026-03-21T04:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.878798 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.887702 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cc87971f-e8fc-454d-8513-957a0bbad389-serviceca\") pod \"node-ca-kr988\" (UID: \"cc87971f-e8fc-454d-8513-957a0bbad389\") " pod="openshift-image-registry/node-ca-kr988" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.887754 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2kj8\" (UniqueName: \"kubernetes.io/projected/cc87971f-e8fc-454d-8513-957a0bbad389-kube-api-access-q2kj8\") pod \"node-ca-kr988\" (UID: \"cc87971f-e8fc-454d-8513-957a0bbad389\") " pod="openshift-image-registry/node-ca-kr988" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.887774 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc87971f-e8fc-454d-8513-957a0bbad389-host\") pod \"node-ca-kr988\" (UID: \"cc87971f-e8fc-454d-8513-957a0bbad389\") " pod="openshift-image-registry/node-ca-kr988" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.887823 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc87971f-e8fc-454d-8513-957a0bbad389-host\") pod \"node-ca-kr988\" (UID: \"cc87971f-e8fc-454d-8513-957a0bbad389\") " pod="openshift-image-registry/node-ca-kr988" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.888684 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cc87971f-e8fc-454d-8513-957a0bbad389-serviceca\") pod \"node-ca-kr988\" (UID: \"cc87971f-e8fc-454d-8513-957a0bbad389\") " pod="openshift-image-registry/node-ca-kr988" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.895562 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.904625 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.911757 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.918613 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2kj8\" (UniqueName: \"kubernetes.io/projected/cc87971f-e8fc-454d-8513-957a0bbad389-kube-api-access-q2kj8\") pod \"node-ca-kr988\" (UID: \"cc87971f-e8fc-454d-8513-957a0bbad389\") " pod="openshift-image-registry/node-ca-kr988" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.922357 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.936713 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.948239 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.958775 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.967832 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.973842 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.973880 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.973892 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.973909 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:04 crc kubenswrapper[4775]: I0321 04:49:04.973921 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:04Z","lastTransitionTime":"2026-03-21T04:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.059148 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-kr988" Mar 21 04:49:05 crc kubenswrapper[4775]: W0321 04:49:05.070487 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc87971f_e8fc_454d_8513_957a0bbad389.slice/crio-45d54ab10f38d7c9c90f7ae9127338cd3e665ac080edc58a170dfeae4b30c23b WatchSource:0}: Error finding container 45d54ab10f38d7c9c90f7ae9127338cd3e665ac080edc58a170dfeae4b30c23b: Status 404 returned error can't find the container with id 45d54ab10f38d7c9c90f7ae9127338cd3e665ac080edc58a170dfeae4b30c23b Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.083472 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.083517 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.083529 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.083552 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.083564 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:05Z","lastTransitionTime":"2026-03-21T04:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.185506 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.185556 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.185568 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.185587 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.185599 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:05Z","lastTransitionTime":"2026-03-21T04:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.288825 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.288868 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.288880 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.288898 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.288914 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:05Z","lastTransitionTime":"2026-03-21T04:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.391522 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.391547 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.391555 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.391568 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.391577 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:05Z","lastTransitionTime":"2026-03-21T04:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.494068 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.494143 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.494156 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.494174 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.494186 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:05Z","lastTransitionTime":"2026-03-21T04:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.596277 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.596311 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.596320 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.596333 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.596342 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:05Z","lastTransitionTime":"2026-03-21T04:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.648249 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0"} Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.648298 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd"} Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.654068 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerStarted","Data":"5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99"} Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.654131 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerStarted","Data":"f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28"} Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.654145 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerStarted","Data":"70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4"} Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.655245 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-kr988" event={"ID":"cc87971f-e8fc-454d-8513-957a0bbad389","Type":"ContainerStarted","Data":"88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8"} Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.655274 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-kr988" event={"ID":"cc87971f-e8fc-454d-8513-957a0bbad389","Type":"ContainerStarted","Data":"45d54ab10f38d7c9c90f7ae9127338cd3e665ac080edc58a170dfeae4b30c23b"} Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.657011 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" event={"ID":"957cde70-ca20-438a-a4bf-42481dddb2db","Type":"ContainerStarted","Data":"54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7"} Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.661445 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.661500 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: E0321 04:49:05.661584 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.672620 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.681715 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.692942 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.697975 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.698008 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.698017 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.698031 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.698041 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:05Z","lastTransitionTime":"2026-03-21T04:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.710925 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.721362 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.736482 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.748000 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.759331 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.770099 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.781175 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.789088 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.798379 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.799827 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.799854 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.799863 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.799877 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.799899 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:05Z","lastTransitionTime":"2026-03-21T04:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.810861 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.819087 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.832081 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.843296 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.853303 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.862581 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.874744 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.896768 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.902226 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.902278 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.902290 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.902311 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.902329 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:05Z","lastTransitionTime":"2026-03-21T04:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.910312 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.929218 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.941156 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.950921 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.958845 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.967852 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:05 crc kubenswrapper[4775]: I0321 04:49:05.975355 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.005224 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.005282 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.005295 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.005317 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.005331 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:06Z","lastTransitionTime":"2026-03-21T04:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.112323 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.112876 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.112889 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.112910 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.112922 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:06Z","lastTransitionTime":"2026-03-21T04:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.215584 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.215628 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.215640 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.215657 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.215668 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:06Z","lastTransitionTime":"2026-03-21T04:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.318361 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.318418 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.318428 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.318447 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.318458 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:06Z","lastTransitionTime":"2026-03-21T04:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.421130 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.421182 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.421194 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.421212 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.421406 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:06Z","lastTransitionTime":"2026-03-21T04:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.524284 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.524323 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.524340 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.524360 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.524374 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:06Z","lastTransitionTime":"2026-03-21T04:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.627396 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.627873 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.627887 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.627906 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.627918 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:06Z","lastTransitionTime":"2026-03-21T04:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.660433 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:06 crc kubenswrapper[4775]: E0321 04:49:06.660570 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.660501 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:06 crc kubenswrapper[4775]: E0321 04:49:06.660904 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.665789 4775 generic.go:334] "Generic (PLEG): container finished" podID="957cde70-ca20-438a-a4bf-42481dddb2db" containerID="54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7" exitCode=0 Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.665830 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" event={"ID":"957cde70-ca20-438a-a4bf-42481dddb2db","Type":"ContainerDied","Data":"54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7"} Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.685439 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.698525 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.718969 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.730687 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.730965 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.730996 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.731004 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.731018 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.731027 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:06Z","lastTransitionTime":"2026-03-21T04:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.743982 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.756561 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.769754 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.782751 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.793352 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.808399 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.824900 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.834056 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.834080 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.834089 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.834101 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.834109 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:06Z","lastTransitionTime":"2026-03-21T04:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.837929 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.848282 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.863307 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.937148 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.937192 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.937262 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.937280 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:06 crc kubenswrapper[4775]: I0321 04:49:06.937292 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:06Z","lastTransitionTime":"2026-03-21T04:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.039565 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.039606 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.039618 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.039638 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.039650 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:07Z","lastTransitionTime":"2026-03-21T04:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.142210 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.142251 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.142263 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.142279 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.142291 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:07Z","lastTransitionTime":"2026-03-21T04:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.244596 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.244645 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.244658 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.244675 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.244689 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:07Z","lastTransitionTime":"2026-03-21T04:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.347229 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.347277 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.347287 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.347305 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.347317 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:07Z","lastTransitionTime":"2026-03-21T04:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.449795 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.449852 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.449864 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.449886 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.449896 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:07Z","lastTransitionTime":"2026-03-21T04:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.552857 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.552889 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.552897 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.552911 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.552921 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:07Z","lastTransitionTime":"2026-03-21T04:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.655683 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.655728 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.655737 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.655750 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.655759 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:07Z","lastTransitionTime":"2026-03-21T04:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.661337 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:07 crc kubenswrapper[4775]: E0321 04:49:07.661459 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.671747 4775 generic.go:334] "Generic (PLEG): container finished" podID="957cde70-ca20-438a-a4bf-42481dddb2db" containerID="f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785" exitCode=0 Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.671808 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" event={"ID":"957cde70-ca20-438a-a4bf-42481dddb2db","Type":"ContainerDied","Data":"f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785"} Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.674320 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84"} Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.679075 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.693318 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.702965 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.714448 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.723746 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.733491 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.748715 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.760331 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.760364 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.760372 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.760385 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.760393 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:07Z","lastTransitionTime":"2026-03-21T04:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.762023 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.773545 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.785304 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.795626 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.813827 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.826065 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.853780 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.862341 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.862492 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.862568 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.862633 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.862700 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:07Z","lastTransitionTime":"2026-03-21T04:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.871518 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.887079 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.903787 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.915835 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.928210 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.940963 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.953388 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.963770 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.965916 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.966019 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.966076 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.966152 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.966230 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:07Z","lastTransitionTime":"2026-03-21T04:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.974541 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:07 crc kubenswrapper[4775]: I0321 04:49:07.988302 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.001814 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.012616 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.024328 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.034777 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.069138 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.069409 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.069506 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.069605 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.069682 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:08Z","lastTransitionTime":"2026-03-21T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.167680 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.167731 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.167743 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.167762 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.167773 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:08Z","lastTransitionTime":"2026-03-21T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.181216 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.184570 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.184619 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.184636 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.184656 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.184669 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:08Z","lastTransitionTime":"2026-03-21T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.209953 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.213763 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.213803 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.213815 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.213831 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.213844 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:08Z","lastTransitionTime":"2026-03-21T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.229674 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.233039 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.233066 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.233075 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.233089 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.233098 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:08Z","lastTransitionTime":"2026-03-21T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.245952 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.249004 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.249035 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.249052 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.249067 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.249078 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:08Z","lastTransitionTime":"2026-03-21T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.260135 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.260296 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.262092 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.262149 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.262161 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.262180 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.262191 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:08Z","lastTransitionTime":"2026-03-21T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.323817 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.323926 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.323951 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.324040 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:49:24.324009027 +0000 UTC m=+117.300472661 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.324052 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.324145 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.324206 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 04:49:24.324194433 +0000 UTC m=+117.300658107 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.324257 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 04:49:24.324244614 +0000 UTC m=+117.300708238 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.363915 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.363961 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.363974 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.363993 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.364006 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:08Z","lastTransitionTime":"2026-03-21T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.424975 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.425021 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.425170 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.425187 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.425200 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.425250 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-21 04:49:24.425236455 +0000 UTC m=+117.401700079 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.425277 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.425329 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.425343 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.425408 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-21 04:49:24.425389439 +0000 UTC m=+117.401853123 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.466785 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.466822 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.466833 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.466849 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.466860 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:08Z","lastTransitionTime":"2026-03-21T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.569220 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.569518 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.569536 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.569554 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.569566 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:08Z","lastTransitionTime":"2026-03-21T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.660604 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.660632 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.660744 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:08 crc kubenswrapper[4775]: E0321 04:49:08.661142 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.671542 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.671576 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.671587 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.671611 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.671624 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:08Z","lastTransitionTime":"2026-03-21T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.681428 4775 generic.go:334] "Generic (PLEG): container finished" podID="957cde70-ca20-438a-a4bf-42481dddb2db" containerID="362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5" exitCode=0 Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.681496 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" event={"ID":"957cde70-ca20-438a-a4bf-42481dddb2db","Type":"ContainerDied","Data":"362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5"} Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.685140 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerStarted","Data":"2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d"} Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.692007 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.702081 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.715479 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.726575 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.738003 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.755460 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.772640 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.773603 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.773626 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.773634 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.773647 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.773656 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:08Z","lastTransitionTime":"2026-03-21T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.784296 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.795151 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.805211 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.814072 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.824639 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.838073 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.848448 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:08Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.875847 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.875879 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.875888 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.875901 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.875911 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:08Z","lastTransitionTime":"2026-03-21T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.978617 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.978661 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.978674 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.978692 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:08 crc kubenswrapper[4775]: I0321 04:49:08.978705 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:08Z","lastTransitionTime":"2026-03-21T04:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.081213 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.081255 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.081263 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.081294 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.081306 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:09Z","lastTransitionTime":"2026-03-21T04:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.184546 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.184602 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.184615 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.184634 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.184647 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:09Z","lastTransitionTime":"2026-03-21T04:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.286523 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.286558 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.286567 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.286580 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.286591 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:09Z","lastTransitionTime":"2026-03-21T04:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.389491 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.389543 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.389557 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.389572 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.389586 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:09Z","lastTransitionTime":"2026-03-21T04:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.491966 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.492040 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.492060 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.492091 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.492111 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:09Z","lastTransitionTime":"2026-03-21T04:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.594802 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.594869 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.594880 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.594894 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.594903 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:09Z","lastTransitionTime":"2026-03-21T04:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.661046 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:09 crc kubenswrapper[4775]: E0321 04:49:09.661196 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.693228 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerStarted","Data":"55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4"} Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.694603 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.694750 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.694806 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.698583 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.698623 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.698639 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.698659 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.698674 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:09Z","lastTransitionTime":"2026-03-21T04:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.710499 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.726409 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.736158 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.738945 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.743004 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.755580 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.776845 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.790680 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.801434 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.801470 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.801491 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.801506 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.801517 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:09Z","lastTransitionTime":"2026-03-21T04:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.811461 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.826380 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.842275 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.856419 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.870962 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.882345 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.891357 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.903895 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.903938 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.903949 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.903966 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.903978 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:09Z","lastTransitionTime":"2026-03-21T04:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.906053 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.917922 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.930246 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.947541 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.962253 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:09 crc kubenswrapper[4775]: I0321 04:49:09.979313 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.000188 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.006456 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.006784 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.006796 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.006815 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.006825 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:10Z","lastTransitionTime":"2026-03-21T04:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.019653 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.033160 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.044213 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.055846 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.066807 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.079810 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.096831 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.108604 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.109147 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.109205 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.109242 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.109263 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.109275 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:10Z","lastTransitionTime":"2026-03-21T04:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.211033 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.211084 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.211095 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.211110 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.211137 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:10Z","lastTransitionTime":"2026-03-21T04:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.313168 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.313210 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.313220 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.313236 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.313246 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:10Z","lastTransitionTime":"2026-03-21T04:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.415313 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.415355 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.415367 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.415387 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.415399 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:10Z","lastTransitionTime":"2026-03-21T04:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.517627 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.517663 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.517672 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.517686 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.517696 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:10Z","lastTransitionTime":"2026-03-21T04:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.621274 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.621348 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.621370 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.621393 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.621411 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:10Z","lastTransitionTime":"2026-03-21T04:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.661305 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.661396 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:10 crc kubenswrapper[4775]: E0321 04:49:10.661539 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:10 crc kubenswrapper[4775]: E0321 04:49:10.661733 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.697305 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143"} Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.707697 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" event={"ID":"957cde70-ca20-438a-a4bf-42481dddb2db","Type":"ContainerStarted","Data":"56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd"} Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.710289 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.723150 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.724205 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.724277 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.724288 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.724301 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.724312 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:10Z","lastTransitionTime":"2026-03-21T04:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.735947 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.748265 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.771048 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.782388 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.798202 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.808331 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.817169 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.826802 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.826841 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.826849 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.826868 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.826877 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:10Z","lastTransitionTime":"2026-03-21T04:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.833828 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.846281 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.858051 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.868673 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.881422 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.898935 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.913938 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.929761 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.929804 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.929818 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.929833 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.929845 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:10Z","lastTransitionTime":"2026-03-21T04:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.934072 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.959828 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.976809 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:10 crc kubenswrapper[4775]: I0321 04:49:10.999451 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:10Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.011728 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:11Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.023111 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:11Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.032529 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.032590 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.032641 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.032657 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.032665 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:11Z","lastTransitionTime":"2026-03-21T04:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.038817 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:11Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.054479 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:11Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.070021 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:11Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.082751 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:11Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.094481 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:11Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.107873 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:11Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.134383 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.134421 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.134429 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.134445 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.134466 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:11Z","lastTransitionTime":"2026-03-21T04:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.236889 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.236922 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.236931 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.236944 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.236954 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:11Z","lastTransitionTime":"2026-03-21T04:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.339525 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.339575 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.339586 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.339601 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.339946 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:11Z","lastTransitionTime":"2026-03-21T04:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.443139 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.443192 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.443207 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.443227 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.443241 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:11Z","lastTransitionTime":"2026-03-21T04:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.544880 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.544918 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.544930 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.544945 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.544955 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:11Z","lastTransitionTime":"2026-03-21T04:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.647643 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.647677 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.647686 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.647699 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.647709 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:11Z","lastTransitionTime":"2026-03-21T04:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.661411 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:11 crc kubenswrapper[4775]: E0321 04:49:11.661566 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.749546 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.749572 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.749580 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.749593 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.749602 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:11Z","lastTransitionTime":"2026-03-21T04:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.852405 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.852764 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.852774 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.852789 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.852799 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:11Z","lastTransitionTime":"2026-03-21T04:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.955149 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.955210 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.955230 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.955255 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:11 crc kubenswrapper[4775]: I0321 04:49:11.955269 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:11Z","lastTransitionTime":"2026-03-21T04:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.057889 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.057926 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.057935 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.057947 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.057955 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:12Z","lastTransitionTime":"2026-03-21T04:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.159999 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.160041 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.160056 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.160072 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.160084 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:12Z","lastTransitionTime":"2026-03-21T04:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.262403 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.262444 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.262459 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.262476 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.262488 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:12Z","lastTransitionTime":"2026-03-21T04:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.365385 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.365449 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.365461 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.365481 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.365496 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:12Z","lastTransitionTime":"2026-03-21T04:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.467694 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.467732 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.467740 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.467756 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.467765 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:12Z","lastTransitionTime":"2026-03-21T04:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.570012 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.570061 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.570071 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.570087 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.570098 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:12Z","lastTransitionTime":"2026-03-21T04:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.661316 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.661407 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:12 crc kubenswrapper[4775]: E0321 04:49:12.661513 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:12 crc kubenswrapper[4775]: E0321 04:49:12.661653 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.673038 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.673082 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.673094 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.673106 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:12 crc kubenswrapper[4775]: I0321 04:49:12.673128 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:12Z","lastTransitionTime":"2026-03-21T04:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.744978 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:13 crc kubenswrapper[4775]: E0321 04:49:13.745136 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.748846 4775 generic.go:334] "Generic (PLEG): container finished" podID="957cde70-ca20-438a-a4bf-42481dddb2db" containerID="56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd" exitCode=0 Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.748940 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.749039 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:13 crc kubenswrapper[4775]: E0321 04:49:13.749053 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.749065 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.749075 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.749096 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.749106 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:13Z","lastTransitionTime":"2026-03-21T04:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.761721 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" event={"ID":"957cde70-ca20-438a-a4bf-42481dddb2db","Type":"ContainerDied","Data":"56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd"} Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.788844 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:13Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.804425 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:13Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.830757 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:13Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.846385 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:13Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.851138 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.851174 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.851185 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.851202 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.851214 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:13Z","lastTransitionTime":"2026-03-21T04:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.859914 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:13Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.870347 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:13Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.884078 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:13Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.897322 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:13Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.912861 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:13Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.929145 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:13Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.953614 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.953641 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.953649 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.953661 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.953674 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:13Z","lastTransitionTime":"2026-03-21T04:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.954781 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:13Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.973764 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:13Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.986213 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:13Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:13 crc kubenswrapper[4775]: I0321 04:49:13.997694 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:13Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.057428 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.057457 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.057465 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.057478 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.057487 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:14Z","lastTransitionTime":"2026-03-21T04:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.149957 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz"] Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.150422 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.152287 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.153063 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.160756 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.160791 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.160801 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.160816 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.160827 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:14Z","lastTransitionTime":"2026-03-21T04:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.175256 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.188339 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.198858 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.212154 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.221620 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.232825 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.247375 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.249589 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/38cb395c-744c-4c2e-9e32-b6cb206a9c5b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-62jtz\" (UID: \"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.249754 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/38cb395c-744c-4c2e-9e32-b6cb206a9c5b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-62jtz\" (UID: \"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.249844 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4d4t\" (UniqueName: \"kubernetes.io/projected/38cb395c-744c-4c2e-9e32-b6cb206a9c5b-kube-api-access-l4d4t\") pod \"ovnkube-control-plane-749d76644c-62jtz\" (UID: \"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.249955 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/38cb395c-744c-4c2e-9e32-b6cb206a9c5b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-62jtz\" (UID: \"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.258197 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.262590 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.262634 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.262646 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.262662 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.262673 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:14Z","lastTransitionTime":"2026-03-21T04:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.271968 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.283617 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.294726 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.305256 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.322232 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.347707 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.351644 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/38cb395c-744c-4c2e-9e32-b6cb206a9c5b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-62jtz\" (UID: \"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.351708 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/38cb395c-744c-4c2e-9e32-b6cb206a9c5b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-62jtz\" (UID: \"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.351729 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4d4t\" (UniqueName: \"kubernetes.io/projected/38cb395c-744c-4c2e-9e32-b6cb206a9c5b-kube-api-access-l4d4t\") pod \"ovnkube-control-plane-749d76644c-62jtz\" (UID: \"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.351750 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/38cb395c-744c-4c2e-9e32-b6cb206a9c5b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-62jtz\" (UID: \"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.352934 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/38cb395c-744c-4c2e-9e32-b6cb206a9c5b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-62jtz\" (UID: \"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.353251 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/38cb395c-744c-4c2e-9e32-b6cb206a9c5b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-62jtz\" (UID: \"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.359899 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/38cb395c-744c-4c2e-9e32-b6cb206a9c5b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-62jtz\" (UID: \"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.364810 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.364843 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.364853 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.364868 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.364878 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:14Z","lastTransitionTime":"2026-03-21T04:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.371675 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4d4t\" (UniqueName: \"kubernetes.io/projected/38cb395c-744c-4c2e-9e32-b6cb206a9c5b-kube-api-access-l4d4t\") pod \"ovnkube-control-plane-749d76644c-62jtz\" (UID: \"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.379392 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.467421 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.467624 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.467766 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.467909 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.468049 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:14Z","lastTransitionTime":"2026-03-21T04:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.475610 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" Mar 21 04:49:14 crc kubenswrapper[4775]: W0321 04:49:14.510326 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38cb395c_744c_4c2e_9e32_b6cb206a9c5b.slice/crio-76f13cfa83a1aa67a2dd389b280570ab2bbbdd5b23c63625d56e8d41b50a46d4 WatchSource:0}: Error finding container 76f13cfa83a1aa67a2dd389b280570ab2bbbdd5b23c63625d56e8d41b50a46d4: Status 404 returned error can't find the container with id 76f13cfa83a1aa67a2dd389b280570ab2bbbdd5b23c63625d56e8d41b50a46d4 Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.575331 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.575367 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.575377 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.575392 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.575401 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:14Z","lastTransitionTime":"2026-03-21T04:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.661393 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:14 crc kubenswrapper[4775]: E0321 04:49:14.661734 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.661799 4775 scope.go:117] "RemoveContainer" containerID="85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.677436 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.677481 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.677491 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.677508 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.677519 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:14Z","lastTransitionTime":"2026-03-21T04:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.754207 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" event={"ID":"957cde70-ca20-438a-a4bf-42481dddb2db","Type":"ContainerStarted","Data":"b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1"} Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.755973 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" event={"ID":"38cb395c-744c-4c2e-9e32-b6cb206a9c5b","Type":"ContainerStarted","Data":"76f13cfa83a1aa67a2dd389b280570ab2bbbdd5b23c63625d56e8d41b50a46d4"} Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.766934 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.780545 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.780570 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.780593 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.780607 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.780616 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:14Z","lastTransitionTime":"2026-03-21T04:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.781157 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.792697 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.806150 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.823681 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.838407 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.863589 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.875039 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.882576 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.882608 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.882618 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.882635 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.882647 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:14Z","lastTransitionTime":"2026-03-21T04:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.885111 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.898237 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.909246 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.920832 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.929108 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.942699 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.954625 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:14Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.984944 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.984979 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.984988 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.985003 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:14 crc kubenswrapper[4775]: I0321 04:49:14.985011 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:14Z","lastTransitionTime":"2026-03-21T04:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.087574 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.087613 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.087622 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.087637 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.087647 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:15Z","lastTransitionTime":"2026-03-21T04:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.190102 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.190153 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.190165 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.190179 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.190190 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:15Z","lastTransitionTime":"2026-03-21T04:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.293005 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.293039 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.293050 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.293067 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.293078 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:15Z","lastTransitionTime":"2026-03-21T04:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.395976 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.396030 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.396043 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.396060 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.396072 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:15Z","lastTransitionTime":"2026-03-21T04:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.499034 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.499076 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.499089 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.499108 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.499135 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:15Z","lastTransitionTime":"2026-03-21T04:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.601531 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.601567 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.601577 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.601592 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.601601 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:15Z","lastTransitionTime":"2026-03-21T04:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.661069 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.661248 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:15 crc kubenswrapper[4775]: E0321 04:49:15.661600 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:15 crc kubenswrapper[4775]: E0321 04:49:15.661745 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.703915 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.703963 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.703980 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.704023 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.704040 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:15Z","lastTransitionTime":"2026-03-21T04:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.763341 4775 generic.go:334] "Generic (PLEG): container finished" podID="957cde70-ca20-438a-a4bf-42481dddb2db" containerID="b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1" exitCode=0 Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.763384 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" event={"ID":"957cde70-ca20-438a-a4bf-42481dddb2db","Type":"ContainerDied","Data":"b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1"} Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.790032 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:15Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.806111 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:15Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.809485 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.809535 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.809546 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.809564 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.809574 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:15Z","lastTransitionTime":"2026-03-21T04:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.831346 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:15Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.848477 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:15Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.864356 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:15Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.878580 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:15Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.894135 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:15Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.904488 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:15Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.911739 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.911776 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.911787 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.911802 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.911813 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:15Z","lastTransitionTime":"2026-03-21T04:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.915179 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:15Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.929698 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:15Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.941183 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:15Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.954617 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:15Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.968156 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:15Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.978519 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:15Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.990232 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:15Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.992036 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-xk9f5"] Mar 21 04:49:15 crc kubenswrapper[4775]: I0321 04:49:15.992567 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:15 crc kubenswrapper[4775]: E0321 04:49:15.992619 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.001580 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.014424 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.014939 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.015026 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.015040 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.015063 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.015074 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:16Z","lastTransitionTime":"2026-03-21T04:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.025021 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.036463 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.047406 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.057810 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.067323 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.082029 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h68dp\" (UniqueName: \"kubernetes.io/projected/6920413a-2c51-466d-a16e-d14489ae0c6c-kube-api-access-h68dp\") pod \"network-metrics-daemon-xk9f5\" (UID: \"6920413a-2c51-466d-a16e-d14489ae0c6c\") " pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.082067 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs\") pod \"network-metrics-daemon-xk9f5\" (UID: \"6920413a-2c51-466d-a16e-d14489ae0c6c\") " pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.083522 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.093764 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.111633 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.116727 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.116756 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.116763 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.116776 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.116785 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:16Z","lastTransitionTime":"2026-03-21T04:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.122853 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.131515 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.143105 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.154742 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.167934 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.179696 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.183004 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h68dp\" (UniqueName: \"kubernetes.io/projected/6920413a-2c51-466d-a16e-d14489ae0c6c-kube-api-access-h68dp\") pod \"network-metrics-daemon-xk9f5\" (UID: \"6920413a-2c51-466d-a16e-d14489ae0c6c\") " pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.183045 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs\") pod \"network-metrics-daemon-xk9f5\" (UID: \"6920413a-2c51-466d-a16e-d14489ae0c6c\") " pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:16 crc kubenswrapper[4775]: E0321 04:49:16.183278 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 04:49:16 crc kubenswrapper[4775]: E0321 04:49:16.183363 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs podName:6920413a-2c51-466d-a16e-d14489ae0c6c nodeName:}" failed. No retries permitted until 2026-03-21 04:49:16.683343475 +0000 UTC m=+109.659807099 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs") pod "network-metrics-daemon-xk9f5" (UID: "6920413a-2c51-466d-a16e-d14489ae0c6c") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.202137 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h68dp\" (UniqueName: \"kubernetes.io/projected/6920413a-2c51-466d-a16e-d14489ae0c6c-kube-api-access-h68dp\") pod \"network-metrics-daemon-xk9f5\" (UID: \"6920413a-2c51-466d-a16e-d14489ae0c6c\") " pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.219872 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.219904 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.219912 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.219925 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.219933 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:16Z","lastTransitionTime":"2026-03-21T04:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.323058 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.323094 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.323104 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.323133 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.323144 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:16Z","lastTransitionTime":"2026-03-21T04:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.425629 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.425661 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.425672 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.425687 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.425699 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:16Z","lastTransitionTime":"2026-03-21T04:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.528254 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.528289 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.528299 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.528315 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.528326 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:16Z","lastTransitionTime":"2026-03-21T04:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.630816 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.630852 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.630863 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.630880 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.630889 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:16Z","lastTransitionTime":"2026-03-21T04:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.660464 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:16 crc kubenswrapper[4775]: E0321 04:49:16.660577 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.687700 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs\") pod \"network-metrics-daemon-xk9f5\" (UID: \"6920413a-2c51-466d-a16e-d14489ae0c6c\") " pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:16 crc kubenswrapper[4775]: E0321 04:49:16.687831 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 04:49:16 crc kubenswrapper[4775]: E0321 04:49:16.687899 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs podName:6920413a-2c51-466d-a16e-d14489ae0c6c nodeName:}" failed. No retries permitted until 2026-03-21 04:49:17.687880058 +0000 UTC m=+110.664343682 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs") pod "network-metrics-daemon-xk9f5" (UID: "6920413a-2c51-466d-a16e-d14489ae0c6c") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.733053 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.733100 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.733110 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.733157 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.733169 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:16Z","lastTransitionTime":"2026-03-21T04:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.768156 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.769379 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401"} Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.770577 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.772626 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" event={"ID":"38cb395c-744c-4c2e-9e32-b6cb206a9c5b","Type":"ContainerStarted","Data":"226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235"} Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.772680 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" event={"ID":"38cb395c-744c-4c2e-9e32-b6cb206a9c5b","Type":"ContainerStarted","Data":"99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3"} Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.783836 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" event={"ID":"957cde70-ca20-438a-a4bf-42481dddb2db","Type":"ContainerStarted","Data":"12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674"} Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.786035 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzqtk_a69d31f5-deeb-4860-be96-ed5547831685/ovnkube-controller/0.log" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.788493 4775 generic.go:334] "Generic (PLEG): container finished" podID="a69d31f5-deeb-4860-be96-ed5547831685" containerID="55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4" exitCode=1 Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.788534 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerDied","Data":"55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4"} Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.789249 4775 scope.go:117] "RemoveContainer" containerID="55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.804292 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.816416 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.834792 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.835100 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.835133 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.835145 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.835158 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.835166 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:16Z","lastTransitionTime":"2026-03-21T04:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.848683 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.859926 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.874421 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.886281 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.897202 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.911145 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.921279 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.936470 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.937667 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.937698 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.937708 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.937724 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.937732 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:16Z","lastTransitionTime":"2026-03-21T04:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.948051 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.961531 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.976935 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:16 crc kubenswrapper[4775]: I0321 04:49:16.990922 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:16Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.004762 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.028477 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.040229 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.040268 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.040279 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.040295 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.040305 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:17Z","lastTransitionTime":"2026-03-21T04:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.045995 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.058042 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.073275 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.084349 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.101828 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.115636 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.127182 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.142018 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.142108 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.142140 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.142161 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.142175 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:17Z","lastTransitionTime":"2026-03-21T04:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.144338 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"message\\\":\\\"factory.go:160\\\\nI0321 04:49:15.347720 6490 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0321 04:49:15.348031 6490 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0321 04:49:15.348133 6490 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0321 04:49:15.348148 6490 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0321 04:49:15.348174 6490 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0321 04:49:15.348468 6490 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0321 04:49:15.348502 6490 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0321 04:49:15.348771 6490 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0321 04:49:15.348774 6490 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0321 04:49:15.348797 6490 factory.go:656] Stopping watch factory\\\\nI0321 04:49:15.348808 6490 ovnkube.go:599] Stopped ovnkube\\\\nI0321 04:49:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.165887 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.183218 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.198299 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.211298 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.222086 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.235103 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.244690 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.244719 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.244730 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.244745 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.244756 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:17Z","lastTransitionTime":"2026-03-21T04:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.247900 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.347643 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.347714 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.347726 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.347743 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.347756 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:17Z","lastTransitionTime":"2026-03-21T04:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.450208 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.450245 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.450254 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.450268 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.450276 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:17Z","lastTransitionTime":"2026-03-21T04:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.553235 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.553273 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.553282 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.553296 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.553305 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:17Z","lastTransitionTime":"2026-03-21T04:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.659333 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.659376 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.659388 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.659403 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.659413 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:17Z","lastTransitionTime":"2026-03-21T04:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.660671 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.660717 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.660747 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:17 crc kubenswrapper[4775]: E0321 04:49:17.660849 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:17 crc kubenswrapper[4775]: E0321 04:49:17.660933 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:17 crc kubenswrapper[4775]: E0321 04:49:17.661042 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.676618 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.690187 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.701357 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs\") pod \"network-metrics-daemon-xk9f5\" (UID: \"6920413a-2c51-466d-a16e-d14489ae0c6c\") " pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:17 crc kubenswrapper[4775]: E0321 04:49:17.701599 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 04:49:17 crc kubenswrapper[4775]: E0321 04:49:17.701710 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs podName:6920413a-2c51-466d-a16e-d14489ae0c6c nodeName:}" failed. No retries permitted until 2026-03-21 04:49:19.701682356 +0000 UTC m=+112.678146020 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs") pod "network-metrics-daemon-xk9f5" (UID: "6920413a-2c51-466d-a16e-d14489ae0c6c") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.703273 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.713349 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.734163 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.750232 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.761608 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.761649 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.761664 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.761681 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.761723 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:17Z","lastTransitionTime":"2026-03-21T04:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.770707 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"message\\\":\\\"factory.go:160\\\\nI0321 04:49:15.347720 6490 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0321 04:49:15.348031 6490 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0321 04:49:15.348133 6490 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0321 04:49:15.348148 6490 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0321 04:49:15.348174 6490 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0321 04:49:15.348468 6490 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0321 04:49:15.348502 6490 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0321 04:49:15.348771 6490 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0321 04:49:15.348774 6490 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0321 04:49:15.348797 6490 factory.go:656] Stopping watch factory\\\\nI0321 04:49:15.348808 6490 ovnkube.go:599] Stopped ovnkube\\\\nI0321 04:49:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.783924 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.793033 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzqtk_a69d31f5-deeb-4860-be96-ed5547831685/ovnkube-controller/0.log" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.796195 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerStarted","Data":"0286b9718372015f0e06759103c765f3cf62e15f7e0bf343897571c50222731e"} Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.796577 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.808898 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.823954 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.833380 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.843237 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.853775 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.864094 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.864142 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.864162 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.864179 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.864189 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:17Z","lastTransitionTime":"2026-03-21T04:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.868195 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.882786 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.895412 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.910814 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.926847 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.940635 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.954624 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.965850 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.967322 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.967357 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.967366 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.967379 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.967389 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:17Z","lastTransitionTime":"2026-03-21T04:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.976756 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:17 crc kubenswrapper[4775]: I0321 04:49:17.990510 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.001708 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.015508 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.028893 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.040935 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.052860 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.071288 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.071339 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.071351 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.071366 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.071381 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:18Z","lastTransitionTime":"2026-03-21T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.074632 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.089269 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.110055 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0286b9718372015f0e06759103c765f3cf62e15f7e0bf343897571c50222731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"message\\\":\\\"factory.go:160\\\\nI0321 04:49:15.347720 6490 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0321 04:49:15.348031 6490 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0321 04:49:15.348133 6490 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0321 04:49:15.348148 6490 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0321 04:49:15.348174 6490 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0321 04:49:15.348468 6490 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0321 04:49:15.348502 6490 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0321 04:49:15.348771 6490 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0321 04:49:15.348774 6490 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0321 04:49:15.348797 6490 factory.go:656] Stopping watch factory\\\\nI0321 04:49:15.348808 6490 ovnkube.go:599] Stopped ovnkube\\\\nI0321 04:49:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.173542 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.173581 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.173594 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.173611 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.173623 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:18Z","lastTransitionTime":"2026-03-21T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.276025 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.276055 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.276064 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.276080 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.276089 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:18Z","lastTransitionTime":"2026-03-21T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.307468 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.307512 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.307527 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.307545 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.307558 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:18Z","lastTransitionTime":"2026-03-21T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:18 crc kubenswrapper[4775]: E0321 04:49:18.319319 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.322742 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.322780 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.322792 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.322812 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.322824 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:18Z","lastTransitionTime":"2026-03-21T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:18 crc kubenswrapper[4775]: E0321 04:49:18.335913 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.339584 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.339620 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.339629 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.339655 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.339665 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:18Z","lastTransitionTime":"2026-03-21T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:18 crc kubenswrapper[4775]: E0321 04:49:18.352372 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.356001 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.356050 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.356063 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.356087 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.356104 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:18Z","lastTransitionTime":"2026-03-21T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:18 crc kubenswrapper[4775]: E0321 04:49:18.369842 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.373663 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.373704 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.373717 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.373755 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.373768 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:18Z","lastTransitionTime":"2026-03-21T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:18 crc kubenswrapper[4775]: E0321 04:49:18.387244 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: E0321 04:49:18.387412 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.388884 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.388930 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.388942 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.388959 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.388972 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:18Z","lastTransitionTime":"2026-03-21T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.491427 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.491491 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.491506 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.491532 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.491543 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:18Z","lastTransitionTime":"2026-03-21T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.594796 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.594854 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.594867 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.594888 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.594903 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:18Z","lastTransitionTime":"2026-03-21T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.660544 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:18 crc kubenswrapper[4775]: E0321 04:49:18.660679 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.698298 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.698347 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.698357 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.698375 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.698387 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:18Z","lastTransitionTime":"2026-03-21T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.800814 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.800864 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.800877 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.800897 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.800913 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:18Z","lastTransitionTime":"2026-03-21T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.802020 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzqtk_a69d31f5-deeb-4860-be96-ed5547831685/ovnkube-controller/1.log" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.802551 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzqtk_a69d31f5-deeb-4860-be96-ed5547831685/ovnkube-controller/0.log" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.804769 4775 generic.go:334] "Generic (PLEG): container finished" podID="a69d31f5-deeb-4860-be96-ed5547831685" containerID="0286b9718372015f0e06759103c765f3cf62e15f7e0bf343897571c50222731e" exitCode=1 Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.804805 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerDied","Data":"0286b9718372015f0e06759103c765f3cf62e15f7e0bf343897571c50222731e"} Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.804841 4775 scope.go:117] "RemoveContainer" containerID="55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.805465 4775 scope.go:117] "RemoveContainer" containerID="0286b9718372015f0e06759103c765f3cf62e15f7e0bf343897571c50222731e" Mar 21 04:49:18 crc kubenswrapper[4775]: E0321 04:49:18.805697 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mzqtk_openshift-ovn-kubernetes(a69d31f5-deeb-4860-be96-ed5547831685)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" podUID="a69d31f5-deeb-4860-be96-ed5547831685" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.819031 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.829137 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.842668 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.853272 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.865176 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.880692 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.894714 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.903255 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.903304 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.903314 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.903330 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.903343 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:18Z","lastTransitionTime":"2026-03-21T04:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.913435 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.932467 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0286b9718372015f0e06759103c765f3cf62e15f7e0bf343897571c50222731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"message\\\":\\\"factory.go:160\\\\nI0321 04:49:15.347720 6490 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0321 04:49:15.348031 6490 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0321 04:49:15.348133 6490 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0321 04:49:15.348148 6490 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0321 04:49:15.348174 6490 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0321 04:49:15.348468 6490 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0321 04:49:15.348502 6490 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0321 04:49:15.348771 6490 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0321 04:49:15.348774 6490 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0321 04:49:15.348797 6490 factory.go:656] Stopping watch factory\\\\nI0321 04:49:15.348808 6490 ovnkube.go:599] Stopped ovnkube\\\\nI0321 04:49:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0286b9718372015f0e06759103c765f3cf62e15f7e0bf343897571c50222731e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:17Z\\\",\\\"message\\\":\\\"networking-console-plugin-85b44fc459-gdk6g\\\\nI0321 04:49:17.880463 6825 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-khh7x\\\\nI0321 04:49:17.880734 6825 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0321 04:49:17.880494 6825 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI0321 04:49:17.880753 6825 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF0321 04:49:17.880790 6825 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.951034 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.967315 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:18 crc kubenswrapper[4775]: I0321 04:49:18.983477 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:18Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.003222 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:19Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.006080 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.006169 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.006189 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.006213 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.006228 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:19Z","lastTransitionTime":"2026-03-21T04:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.018093 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:19Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.030488 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:19Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.049470 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:19Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.109606 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.109663 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.109676 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.109697 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.109710 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:19Z","lastTransitionTime":"2026-03-21T04:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.212195 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.212238 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.212246 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.212261 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.212272 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:19Z","lastTransitionTime":"2026-03-21T04:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.314502 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.314543 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.314554 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.314568 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.314579 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:19Z","lastTransitionTime":"2026-03-21T04:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.417043 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.417091 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.417147 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.417172 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.417182 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:19Z","lastTransitionTime":"2026-03-21T04:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.519372 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.519409 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.519429 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.519447 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.519457 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:19Z","lastTransitionTime":"2026-03-21T04:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.621559 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.621654 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.621676 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.621701 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.621715 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:19Z","lastTransitionTime":"2026-03-21T04:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.662237 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:19 crc kubenswrapper[4775]: E0321 04:49:19.663032 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.662381 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:19 crc kubenswrapper[4775]: E0321 04:49:19.663139 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.662335 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:19 crc kubenswrapper[4775]: E0321 04:49:19.663198 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.723366 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs\") pod \"network-metrics-daemon-xk9f5\" (UID: \"6920413a-2c51-466d-a16e-d14489ae0c6c\") " pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:19 crc kubenswrapper[4775]: E0321 04:49:19.723480 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 04:49:19 crc kubenswrapper[4775]: E0321 04:49:19.723543 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs podName:6920413a-2c51-466d-a16e-d14489ae0c6c nodeName:}" failed. No retries permitted until 2026-03-21 04:49:23.723526033 +0000 UTC m=+116.699989657 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs") pod "network-metrics-daemon-xk9f5" (UID: "6920413a-2c51-466d-a16e-d14489ae0c6c") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.724927 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.724965 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.724976 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.724993 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.725019 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:19Z","lastTransitionTime":"2026-03-21T04:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.809346 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzqtk_a69d31f5-deeb-4860-be96-ed5547831685/ovnkube-controller/1.log" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.827824 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.827879 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.827893 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.827910 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.827919 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:19Z","lastTransitionTime":"2026-03-21T04:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.930012 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.930048 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.930064 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.930082 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:19 crc kubenswrapper[4775]: I0321 04:49:19.930092 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:19Z","lastTransitionTime":"2026-03-21T04:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.033065 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.033164 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.033179 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.033205 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.033222 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:20Z","lastTransitionTime":"2026-03-21T04:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.135542 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.135603 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.135611 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.135628 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.135638 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:20Z","lastTransitionTime":"2026-03-21T04:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.243318 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.243368 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.243380 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.243398 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.243409 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:20Z","lastTransitionTime":"2026-03-21T04:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.346360 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.346714 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.346727 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.346744 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.346756 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:20Z","lastTransitionTime":"2026-03-21T04:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.449271 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.449313 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.449324 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.449343 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.449356 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:20Z","lastTransitionTime":"2026-03-21T04:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.551618 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.551659 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.551669 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.551684 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.551697 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:20Z","lastTransitionTime":"2026-03-21T04:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.659441 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.659490 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.659501 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.659519 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.659552 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:20Z","lastTransitionTime":"2026-03-21T04:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.660529 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:20 crc kubenswrapper[4775]: E0321 04:49:20.660781 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.762774 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.762816 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.762828 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.762847 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.762859 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:20Z","lastTransitionTime":"2026-03-21T04:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.864888 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.864935 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.864946 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.864962 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.864973 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:20Z","lastTransitionTime":"2026-03-21T04:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.968351 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.968419 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.968439 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.968467 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:20 crc kubenswrapper[4775]: I0321 04:49:20.968488 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:20Z","lastTransitionTime":"2026-03-21T04:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.071665 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.071712 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.071724 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.071741 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.071754 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:21Z","lastTransitionTime":"2026-03-21T04:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.175596 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.175628 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.175637 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.175650 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.175658 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:21Z","lastTransitionTime":"2026-03-21T04:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.278863 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.278915 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.278927 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.278947 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.278960 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:21Z","lastTransitionTime":"2026-03-21T04:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.381307 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.381339 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.381347 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.381360 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.381369 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:21Z","lastTransitionTime":"2026-03-21T04:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.484329 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.484399 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.484417 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.484446 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.484469 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:21Z","lastTransitionTime":"2026-03-21T04:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.586836 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.586881 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.586890 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.586907 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.586923 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:21Z","lastTransitionTime":"2026-03-21T04:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.661149 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.661184 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.661302 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:21 crc kubenswrapper[4775]: E0321 04:49:21.661467 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:21 crc kubenswrapper[4775]: E0321 04:49:21.661648 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:21 crc kubenswrapper[4775]: E0321 04:49:21.661826 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.689521 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.689574 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.689586 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.689604 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.689621 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:21Z","lastTransitionTime":"2026-03-21T04:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.792385 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.792414 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.792425 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.792440 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.792451 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:21Z","lastTransitionTime":"2026-03-21T04:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.894509 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.894546 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.894557 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.894573 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.894583 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:21Z","lastTransitionTime":"2026-03-21T04:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.997191 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.997524 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.997653 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.997784 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:21 crc kubenswrapper[4775]: I0321 04:49:21.997915 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:21Z","lastTransitionTime":"2026-03-21T04:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.099919 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.099984 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.099998 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.100019 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.100030 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:22Z","lastTransitionTime":"2026-03-21T04:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.202239 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.202294 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.202304 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.202320 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.202333 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:22Z","lastTransitionTime":"2026-03-21T04:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.304558 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.304636 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.304651 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.304676 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.304692 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:22Z","lastTransitionTime":"2026-03-21T04:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.408010 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.408079 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.408102 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.408148 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.408161 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:22Z","lastTransitionTime":"2026-03-21T04:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.510651 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.510909 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.511049 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.511263 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.511430 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:22Z","lastTransitionTime":"2026-03-21T04:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.614163 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.614512 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.614661 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.614753 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.614941 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:22Z","lastTransitionTime":"2026-03-21T04:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.661059 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:22 crc kubenswrapper[4775]: E0321 04:49:22.661441 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.716956 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.717001 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.717017 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.717037 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.717052 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:22Z","lastTransitionTime":"2026-03-21T04:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.820431 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.820496 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.820508 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.820530 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.820550 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:22Z","lastTransitionTime":"2026-03-21T04:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.923432 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.923482 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.923491 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.923506 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:22 crc kubenswrapper[4775]: I0321 04:49:22.923515 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:22Z","lastTransitionTime":"2026-03-21T04:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.027151 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.027530 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.027786 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.027967 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.028172 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:23Z","lastTransitionTime":"2026-03-21T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.132217 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.132614 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.132688 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.132774 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.132916 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:23Z","lastTransitionTime":"2026-03-21T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.236395 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.236933 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.237021 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.237180 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.237299 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:23Z","lastTransitionTime":"2026-03-21T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.340342 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.340583 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.340700 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.340801 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.340903 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:23Z","lastTransitionTime":"2026-03-21T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.443794 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.443837 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.443851 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.443866 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.443876 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:23Z","lastTransitionTime":"2026-03-21T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.549756 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.549796 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.549806 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.549821 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.549831 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:23Z","lastTransitionTime":"2026-03-21T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.652640 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.652683 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.652694 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.652709 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.652720 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:23Z","lastTransitionTime":"2026-03-21T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.660981 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.661029 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.661090 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:23 crc kubenswrapper[4775]: E0321 04:49:23.661087 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:23 crc kubenswrapper[4775]: E0321 04:49:23.661184 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:23 crc kubenswrapper[4775]: E0321 04:49:23.661279 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.755089 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.755163 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.755174 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.755193 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.755206 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:23Z","lastTransitionTime":"2026-03-21T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.767645 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs\") pod \"network-metrics-daemon-xk9f5\" (UID: \"6920413a-2c51-466d-a16e-d14489ae0c6c\") " pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:23 crc kubenswrapper[4775]: E0321 04:49:23.767764 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 04:49:23 crc kubenswrapper[4775]: E0321 04:49:23.767817 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs podName:6920413a-2c51-466d-a16e-d14489ae0c6c nodeName:}" failed. No retries permitted until 2026-03-21 04:49:31.767803403 +0000 UTC m=+124.744267027 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs") pod "network-metrics-daemon-xk9f5" (UID: "6920413a-2c51-466d-a16e-d14489ae0c6c") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.857430 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.857463 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.857472 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.857484 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.857493 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:23Z","lastTransitionTime":"2026-03-21T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.959803 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.960255 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.960287 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.960308 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:23 crc kubenswrapper[4775]: I0321 04:49:23.960318 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:23Z","lastTransitionTime":"2026-03-21T04:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.063468 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.063515 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.063526 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.063543 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.063555 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:24Z","lastTransitionTime":"2026-03-21T04:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.166193 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.166239 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.166253 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.166271 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.166282 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:24Z","lastTransitionTime":"2026-03-21T04:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.268507 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.268546 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.268558 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.268574 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.268587 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:24Z","lastTransitionTime":"2026-03-21T04:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.371851 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.371969 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.371994 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:24 crc kubenswrapper[4775]: E0321 04:49:24.372045 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:49:56.372020119 +0000 UTC m=+149.348483763 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:49:24 crc kubenswrapper[4775]: E0321 04:49:24.372058 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 04:49:24 crc kubenswrapper[4775]: E0321 04:49:24.372106 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 04:49:24 crc kubenswrapper[4775]: E0321 04:49:24.372128 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 04:49:56.372103431 +0000 UTC m=+149.348567055 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 04:49:24 crc kubenswrapper[4775]: E0321 04:49:24.372164 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 04:49:56.372151203 +0000 UTC m=+149.348614827 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.372293 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.372306 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.372314 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.372327 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.372335 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:24Z","lastTransitionTime":"2026-03-21T04:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.472864 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.472907 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:24 crc kubenswrapper[4775]: E0321 04:49:24.473002 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 04:49:24 crc kubenswrapper[4775]: E0321 04:49:24.473025 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 04:49:24 crc kubenswrapper[4775]: E0321 04:49:24.473036 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:49:24 crc kubenswrapper[4775]: E0321 04:49:24.473054 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 04:49:24 crc kubenswrapper[4775]: E0321 04:49:24.473072 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 04:49:24 crc kubenswrapper[4775]: E0321 04:49:24.473084 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:49:24 crc kubenswrapper[4775]: E0321 04:49:24.473085 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-21 04:49:56.473069762 +0000 UTC m=+149.449533386 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:49:24 crc kubenswrapper[4775]: E0321 04:49:24.473183 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-21 04:49:56.473169704 +0000 UTC m=+149.449633328 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.474184 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.474297 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.474391 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.474473 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.474551 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:24Z","lastTransitionTime":"2026-03-21T04:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.576841 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.577207 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.577294 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.577383 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.577523 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:24Z","lastTransitionTime":"2026-03-21T04:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.660774 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:24 crc kubenswrapper[4775]: E0321 04:49:24.660903 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.679617 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.679887 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.679968 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.680058 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.680156 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:24Z","lastTransitionTime":"2026-03-21T04:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.783396 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.783438 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.783449 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.783467 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.783479 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:24Z","lastTransitionTime":"2026-03-21T04:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.886094 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.886172 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.886192 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.886214 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.886231 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:24Z","lastTransitionTime":"2026-03-21T04:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.988673 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.989326 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.989396 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.989431 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:24 crc kubenswrapper[4775]: I0321 04:49:24.989454 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:24Z","lastTransitionTime":"2026-03-21T04:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.091856 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.091909 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.091928 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.091951 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.091968 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:25Z","lastTransitionTime":"2026-03-21T04:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.194524 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.194579 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.194607 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.194632 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.194650 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:25Z","lastTransitionTime":"2026-03-21T04:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.297104 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.297167 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.297185 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.297216 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.297232 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:25Z","lastTransitionTime":"2026-03-21T04:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.399181 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.399572 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.399728 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.399830 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.399935 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:25Z","lastTransitionTime":"2026-03-21T04:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.502153 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.502193 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.502212 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.502229 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.502241 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:25Z","lastTransitionTime":"2026-03-21T04:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.604513 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.604770 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.604840 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.604914 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.604987 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:25Z","lastTransitionTime":"2026-03-21T04:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.661364 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:25 crc kubenswrapper[4775]: E0321 04:49:25.661517 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.661597 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.661617 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:25 crc kubenswrapper[4775]: E0321 04:49:25.661753 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:25 crc kubenswrapper[4775]: E0321 04:49:25.661827 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.707420 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.707471 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.707484 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.707503 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.707515 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:25Z","lastTransitionTime":"2026-03-21T04:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.810475 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.810517 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.810528 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.810545 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.810556 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:25Z","lastTransitionTime":"2026-03-21T04:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.912944 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.912990 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.913001 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.913017 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:25 crc kubenswrapper[4775]: I0321 04:49:25.913028 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:25Z","lastTransitionTime":"2026-03-21T04:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.015436 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.015511 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.015522 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.015571 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.015588 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:26Z","lastTransitionTime":"2026-03-21T04:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.117730 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.117775 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.117786 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.117805 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.117817 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:26Z","lastTransitionTime":"2026-03-21T04:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.220298 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.220348 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.220364 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.220386 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.220402 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:26Z","lastTransitionTime":"2026-03-21T04:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.323486 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.323543 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.323565 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.323592 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.323612 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:26Z","lastTransitionTime":"2026-03-21T04:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.426594 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.426666 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.426687 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.426712 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.426729 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:26Z","lastTransitionTime":"2026-03-21T04:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.529296 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.529534 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.529626 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.529718 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.529809 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:26Z","lastTransitionTime":"2026-03-21T04:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.632801 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.633108 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.633240 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.633343 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.633445 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:26Z","lastTransitionTime":"2026-03-21T04:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.661139 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:26 crc kubenswrapper[4775]: E0321 04:49:26.661267 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.735158 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.735199 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.735212 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.735227 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.735241 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:26Z","lastTransitionTime":"2026-03-21T04:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.836556 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.836606 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.836618 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.836632 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.836642 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:26Z","lastTransitionTime":"2026-03-21T04:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.938518 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.938592 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.938604 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.938619 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:26 crc kubenswrapper[4775]: I0321 04:49:26.938630 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:26Z","lastTransitionTime":"2026-03-21T04:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.040730 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.040764 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.040773 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.040787 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.040796 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:27Z","lastTransitionTime":"2026-03-21T04:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.142952 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.142995 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.143008 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.143024 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.143035 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:27Z","lastTransitionTime":"2026-03-21T04:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.246081 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.246152 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.246167 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.246186 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.246197 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:27Z","lastTransitionTime":"2026-03-21T04:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.349295 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.349329 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.349340 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.349355 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.349365 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:27Z","lastTransitionTime":"2026-03-21T04:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.451349 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.451623 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.451699 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.451763 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.451822 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:27Z","lastTransitionTime":"2026-03-21T04:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.554218 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.554463 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.554554 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.554651 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.554742 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:27Z","lastTransitionTime":"2026-03-21T04:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:27 crc kubenswrapper[4775]: E0321 04:49:27.655193 4775 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.660337 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.660420 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:27 crc kubenswrapper[4775]: E0321 04:49:27.660605 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:27 crc kubenswrapper[4775]: E0321 04:49:27.660641 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.660348 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:27 crc kubenswrapper[4775]: E0321 04:49:27.661697 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.678830 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:27Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.698793 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:27Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.710786 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:27Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.726015 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:27Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.746144 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:27Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.762598 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:27Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.785097 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0286b9718372015f0e06759103c765f3cf62e15f7e0bf343897571c50222731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55cbe34c19fee4a7423b106752c0f2fc2f2beebb0b87e92ad58b6ef9fc166db4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"message\\\":\\\"factory.go:160\\\\nI0321 04:49:15.347720 6490 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0321 04:49:15.348031 6490 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0321 04:49:15.348133 6490 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0321 04:49:15.348148 6490 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0321 04:49:15.348174 6490 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0321 04:49:15.348468 6490 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0321 04:49:15.348502 6490 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0321 04:49:15.348771 6490 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0321 04:49:15.348774 6490 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0321 04:49:15.348797 6490 factory.go:656] Stopping watch factory\\\\nI0321 04:49:15.348808 6490 ovnkube.go:599] Stopped ovnkube\\\\nI0321 04:49:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0286b9718372015f0e06759103c765f3cf62e15f7e0bf343897571c50222731e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:17Z\\\",\\\"message\\\":\\\"networking-console-plugin-85b44fc459-gdk6g\\\\nI0321 04:49:17.880463 6825 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-khh7x\\\\nI0321 04:49:17.880734 6825 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0321 04:49:17.880494 6825 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI0321 04:49:17.880753 6825 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF0321 04:49:17.880790 6825 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:27Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.798388 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:27Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.812184 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:27Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.826197 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:27Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.840699 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:27Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.849954 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:27Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.860723 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:27Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.871561 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:27Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.887453 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:27Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:27 crc kubenswrapper[4775]: I0321 04:49:27.898793 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:27Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.661307 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:28 crc kubenswrapper[4775]: E0321 04:49:28.661435 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.745498 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.745543 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.745554 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.745571 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.745585 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:28Z","lastTransitionTime":"2026-03-21T04:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:28 crc kubenswrapper[4775]: E0321 04:49:28.751815 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 04:49:28 crc kubenswrapper[4775]: E0321 04:49:28.759506 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:28Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.763554 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.763587 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.763596 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.763609 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.763618 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:28Z","lastTransitionTime":"2026-03-21T04:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:28 crc kubenswrapper[4775]: E0321 04:49:28.777168 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:28Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.782240 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.782299 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.782318 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.782342 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.782363 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:28Z","lastTransitionTime":"2026-03-21T04:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:28 crc kubenswrapper[4775]: E0321 04:49:28.798278 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:28Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.802342 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.802379 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.802391 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.802406 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.802416 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:28Z","lastTransitionTime":"2026-03-21T04:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:28 crc kubenswrapper[4775]: E0321 04:49:28.816592 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:28Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.820131 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.820173 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.820186 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.820208 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:28 crc kubenswrapper[4775]: I0321 04:49:28.820225 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:28Z","lastTransitionTime":"2026-03-21T04:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:28 crc kubenswrapper[4775]: E0321 04:49:28.831427 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:28Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:28 crc kubenswrapper[4775]: E0321 04:49:28.831565 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 21 04:49:29 crc kubenswrapper[4775]: I0321 04:49:29.660607 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:29 crc kubenswrapper[4775]: I0321 04:49:29.660707 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:29 crc kubenswrapper[4775]: E0321 04:49:29.660736 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:29 crc kubenswrapper[4775]: I0321 04:49:29.660788 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:29 crc kubenswrapper[4775]: E0321 04:49:29.660888 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:29 crc kubenswrapper[4775]: E0321 04:49:29.660994 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:30 crc kubenswrapper[4775]: I0321 04:49:30.661242 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:30 crc kubenswrapper[4775]: E0321 04:49:30.661376 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:31 crc kubenswrapper[4775]: I0321 04:49:31.660784 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:31 crc kubenswrapper[4775]: E0321 04:49:31.660895 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:31 crc kubenswrapper[4775]: I0321 04:49:31.660796 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:31 crc kubenswrapper[4775]: E0321 04:49:31.660961 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:31 crc kubenswrapper[4775]: I0321 04:49:31.661425 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:31 crc kubenswrapper[4775]: E0321 04:49:31.661643 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:31 crc kubenswrapper[4775]: I0321 04:49:31.848220 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs\") pod \"network-metrics-daemon-xk9f5\" (UID: \"6920413a-2c51-466d-a16e-d14489ae0c6c\") " pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:31 crc kubenswrapper[4775]: E0321 04:49:31.848358 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 04:49:31 crc kubenswrapper[4775]: E0321 04:49:31.848511 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs podName:6920413a-2c51-466d-a16e-d14489ae0c6c nodeName:}" failed. No retries permitted until 2026-03-21 04:49:47.848464609 +0000 UTC m=+140.824928233 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs") pod "network-metrics-daemon-xk9f5" (UID: "6920413a-2c51-466d-a16e-d14489ae0c6c") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 04:49:32 crc kubenswrapper[4775]: I0321 04:49:32.661287 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:32 crc kubenswrapper[4775]: E0321 04:49:32.661480 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:32 crc kubenswrapper[4775]: I0321 04:49:32.834449 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:32 crc kubenswrapper[4775]: I0321 04:49:32.836216 4775 scope.go:117] "RemoveContainer" containerID="0286b9718372015f0e06759103c765f3cf62e15f7e0bf343897571c50222731e" Mar 21 04:49:32 crc kubenswrapper[4775]: I0321 04:49:32.863864 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0286b9718372015f0e06759103c765f3cf62e15f7e0bf343897571c50222731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0286b9718372015f0e06759103c765f3cf62e15f7e0bf343897571c50222731e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:17Z\\\",\\\"message\\\":\\\"networking-console-plugin-85b44fc459-gdk6g\\\\nI0321 04:49:17.880463 6825 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-khh7x\\\\nI0321 04:49:17.880734 6825 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0321 04:49:17.880494 6825 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI0321 04:49:17.880753 6825 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF0321 04:49:17.880790 6825 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mzqtk_openshift-ovn-kubernetes(a69d31f5-deeb-4860-be96-ed5547831685)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:32Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:32 crc kubenswrapper[4775]: I0321 04:49:32.890365 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:32Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:32 crc kubenswrapper[4775]: I0321 04:49:32.906264 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:32Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:32 crc kubenswrapper[4775]: I0321 04:49:32.921918 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:32Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:32 crc kubenswrapper[4775]: I0321 04:49:32.943553 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:32Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:32 crc kubenswrapper[4775]: I0321 04:49:32.953963 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:32Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:32 crc kubenswrapper[4775]: I0321 04:49:32.964874 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:32Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:32 crc kubenswrapper[4775]: I0321 04:49:32.977789 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:32Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:32 crc kubenswrapper[4775]: I0321 04:49:32.995345 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:32Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.007948 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.018990 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.033380 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.046504 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.060612 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.074103 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.085723 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.136915 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.155530 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.169324 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.191778 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0286b9718372015f0e06759103c765f3cf62e15f7e0bf343897571c50222731e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0286b9718372015f0e06759103c765f3cf62e15f7e0bf343897571c50222731e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:17Z\\\",\\\"message\\\":\\\"networking-console-plugin-85b44fc459-gdk6g\\\\nI0321 04:49:17.880463 6825 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-khh7x\\\\nI0321 04:49:17.880734 6825 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0321 04:49:17.880494 6825 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI0321 04:49:17.880753 6825 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF0321 04:49:17.880790 6825 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mzqtk_openshift-ovn-kubernetes(a69d31f5-deeb-4860-be96-ed5547831685)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.206665 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.217807 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.229242 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.245416 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.259768 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.274075 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.285695 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.300807 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.312554 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.330467 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.344134 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.356548 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.368362 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.661139 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.661180 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.661223 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:33 crc kubenswrapper[4775]: E0321 04:49:33.661264 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:33 crc kubenswrapper[4775]: E0321 04:49:33.661374 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:33 crc kubenswrapper[4775]: E0321 04:49:33.661453 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:33 crc kubenswrapper[4775]: E0321 04:49:33.753532 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.860774 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzqtk_a69d31f5-deeb-4860-be96-ed5547831685/ovnkube-controller/1.log" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.863769 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerStarted","Data":"a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b"} Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.864249 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.877764 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.894961 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.911700 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.926425 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.940877 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.954817 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.970189 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:33 crc kubenswrapper[4775]: I0321 04:49:33.995335 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:33Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.009585 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:34Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.043475 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0286b9718372015f0e06759103c765f3cf62e15f7e0bf343897571c50222731e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:17Z\\\",\\\"message\\\":\\\"networking-console-plugin-85b44fc459-gdk6g\\\\nI0321 04:49:17.880463 6825 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-khh7x\\\\nI0321 04:49:17.880734 6825 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0321 04:49:17.880494 6825 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI0321 04:49:17.880753 6825 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF0321 04:49:17.880790 6825 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:34Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.056946 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:34Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.071284 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:34Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.084282 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:34Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.101605 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:34Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.114239 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:34Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.127252 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:34Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.660777 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:34 crc kubenswrapper[4775]: E0321 04:49:34.660914 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.868864 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzqtk_a69d31f5-deeb-4860-be96-ed5547831685/ovnkube-controller/2.log" Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.869444 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzqtk_a69d31f5-deeb-4860-be96-ed5547831685/ovnkube-controller/1.log" Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.895377 4775 generic.go:334] "Generic (PLEG): container finished" podID="a69d31f5-deeb-4860-be96-ed5547831685" containerID="a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b" exitCode=1 Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.895432 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerDied","Data":"a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b"} Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.895484 4775 scope.go:117] "RemoveContainer" containerID="0286b9718372015f0e06759103c765f3cf62e15f7e0bf343897571c50222731e" Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.895956 4775 scope.go:117] "RemoveContainer" containerID="a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b" Mar 21 04:49:34 crc kubenswrapper[4775]: E0321 04:49:34.896145 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mzqtk_openshift-ovn-kubernetes(a69d31f5-deeb-4860-be96-ed5547831685)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" podUID="a69d31f5-deeb-4860-be96-ed5547831685" Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.910528 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:34Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.921928 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:34Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.931696 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:34Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.941921 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:34Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.954053 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:34Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.970263 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:34Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:34 crc kubenswrapper[4775]: I0321 04:49:34.985432 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:34Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:35 crc kubenswrapper[4775]: I0321 04:49:35.002396 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:34Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:35 crc kubenswrapper[4775]: I0321 04:49:35.021904 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0286b9718372015f0e06759103c765f3cf62e15f7e0bf343897571c50222731e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:17Z\\\",\\\"message\\\":\\\"networking-console-plugin-85b44fc459-gdk6g\\\\nI0321 04:49:17.880463 6825 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-khh7x\\\\nI0321 04:49:17.880734 6825 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0321 04:49:17.880494 6825 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI0321 04:49:17.880753 6825 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF0321 04:49:17.880790 6825 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:34Z\\\",\\\"message\\\":\\\"eighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.21:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {58a148b3-0a7b-4412-b447-f87788c4883f}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0321 04:49:33.949601 7032 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-etcd-operator/metrics]} name:Service_openshift-etcd-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.188:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {53c717ca-2174-4315-bb03-c937a9c0d9b6}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0321 04:49:33.949663 7032 factory.go:656] Stopping watch factory\\\\nI0321 04:49:33.949677 7032 ovnkube.go:599] Stopped ovnkube\\\\nI0321 04:49:33.949692 7032 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0321 04:49:33.949770 7032 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:35Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:35 crc kubenswrapper[4775]: I0321 04:49:35.040977 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:35Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:35 crc kubenswrapper[4775]: I0321 04:49:35.055667 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:35Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:35 crc kubenswrapper[4775]: I0321 04:49:35.068156 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:35Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:35 crc kubenswrapper[4775]: I0321 04:49:35.080503 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:35Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:35 crc kubenswrapper[4775]: I0321 04:49:35.091389 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:35Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:35 crc kubenswrapper[4775]: I0321 04:49:35.102498 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:35Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:35 crc kubenswrapper[4775]: I0321 04:49:35.118016 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:35Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:35 crc kubenswrapper[4775]: I0321 04:49:35.660680 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:35 crc kubenswrapper[4775]: I0321 04:49:35.660758 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:35 crc kubenswrapper[4775]: E0321 04:49:35.660911 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:35 crc kubenswrapper[4775]: I0321 04:49:35.660959 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:35 crc kubenswrapper[4775]: E0321 04:49:35.661144 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:35 crc kubenswrapper[4775]: E0321 04:49:35.661210 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:35 crc kubenswrapper[4775]: I0321 04:49:35.899454 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzqtk_a69d31f5-deeb-4860-be96-ed5547831685/ovnkube-controller/2.log" Mar 21 04:49:35 crc kubenswrapper[4775]: I0321 04:49:35.902638 4775 scope.go:117] "RemoveContainer" containerID="a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b" Mar 21 04:49:35 crc kubenswrapper[4775]: E0321 04:49:35.902783 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mzqtk_openshift-ovn-kubernetes(a69d31f5-deeb-4860-be96-ed5547831685)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" podUID="a69d31f5-deeb-4860-be96-ed5547831685" Mar 21 04:49:35 crc kubenswrapper[4775]: I0321 04:49:35.915987 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:35Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:35 crc kubenswrapper[4775]: I0321 04:49:35.930332 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:35Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:35 crc kubenswrapper[4775]: I0321 04:49:35.944362 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:35Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:35 crc kubenswrapper[4775]: I0321 04:49:35.956064 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:35Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:35 crc kubenswrapper[4775]: I0321 04:49:35.974431 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:34Z\\\",\\\"message\\\":\\\"eighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.21:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {58a148b3-0a7b-4412-b447-f87788c4883f}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0321 04:49:33.949601 7032 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-etcd-operator/metrics]} name:Service_openshift-etcd-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.188:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {53c717ca-2174-4315-bb03-c937a9c0d9b6}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0321 04:49:33.949663 7032 factory.go:656] Stopping watch factory\\\\nI0321 04:49:33.949677 7032 ovnkube.go:599] Stopped ovnkube\\\\nI0321 04:49:33.949692 7032 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0321 04:49:33.949770 7032 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mzqtk_openshift-ovn-kubernetes(a69d31f5-deeb-4860-be96-ed5547831685)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:35Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:35 crc kubenswrapper[4775]: I0321 04:49:35.994340 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:35Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:36 crc kubenswrapper[4775]: I0321 04:49:36.008639 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:36Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:36 crc kubenswrapper[4775]: I0321 04:49:36.019077 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:36Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:36 crc kubenswrapper[4775]: I0321 04:49:36.030530 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:36Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:36 crc kubenswrapper[4775]: I0321 04:49:36.040161 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:36Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:36 crc kubenswrapper[4775]: I0321 04:49:36.049384 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:36Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:36 crc kubenswrapper[4775]: I0321 04:49:36.060068 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:36Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:36 crc kubenswrapper[4775]: I0321 04:49:36.070579 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:36Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:36 crc kubenswrapper[4775]: I0321 04:49:36.081069 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:36Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:36 crc kubenswrapper[4775]: I0321 04:49:36.093255 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:36Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:36 crc kubenswrapper[4775]: I0321 04:49:36.109518 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:36Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:36 crc kubenswrapper[4775]: I0321 04:49:36.660545 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:36 crc kubenswrapper[4775]: E0321 04:49:36.660679 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:37 crc kubenswrapper[4775]: I0321 04:49:37.660415 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:37 crc kubenswrapper[4775]: I0321 04:49:37.660576 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:37 crc kubenswrapper[4775]: E0321 04:49:37.660814 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:37 crc kubenswrapper[4775]: I0321 04:49:37.660840 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:37 crc kubenswrapper[4775]: E0321 04:49:37.660892 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:37 crc kubenswrapper[4775]: E0321 04:49:37.660965 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:37 crc kubenswrapper[4775]: I0321 04:49:37.679068 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:37Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:37 crc kubenswrapper[4775]: I0321 04:49:37.692437 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:37Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:37 crc kubenswrapper[4775]: I0321 04:49:37.704850 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:37Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:37 crc kubenswrapper[4775]: I0321 04:49:37.716833 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:37Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:37 crc kubenswrapper[4775]: I0321 04:49:37.736806 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:37Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:37 crc kubenswrapper[4775]: I0321 04:49:37.749045 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:37Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:37 crc kubenswrapper[4775]: I0321 04:49:37.772697 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:34Z\\\",\\\"message\\\":\\\"eighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.21:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {58a148b3-0a7b-4412-b447-f87788c4883f}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0321 04:49:33.949601 7032 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-etcd-operator/metrics]} name:Service_openshift-etcd-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.188:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {53c717ca-2174-4315-bb03-c937a9c0d9b6}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0321 04:49:33.949663 7032 factory.go:656] Stopping watch factory\\\\nI0321 04:49:33.949677 7032 ovnkube.go:599] Stopped ovnkube\\\\nI0321 04:49:33.949692 7032 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0321 04:49:33.949770 7032 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mzqtk_openshift-ovn-kubernetes(a69d31f5-deeb-4860-be96-ed5547831685)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:37Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:37 crc kubenswrapper[4775]: I0321 04:49:37.783609 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:37Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:37 crc kubenswrapper[4775]: I0321 04:49:37.795521 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:37Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:37 crc kubenswrapper[4775]: I0321 04:49:37.808021 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:37Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:37 crc kubenswrapper[4775]: I0321 04:49:37.820318 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:37Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:37 crc kubenswrapper[4775]: I0321 04:49:37.841883 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:37Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:37 crc kubenswrapper[4775]: I0321 04:49:37.853268 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:37Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:37 crc kubenswrapper[4775]: I0321 04:49:37.864480 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:37Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:37 crc kubenswrapper[4775]: I0321 04:49:37.880256 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:37Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:37 crc kubenswrapper[4775]: I0321 04:49:37.891953 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:37Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.660320 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:38 crc kubenswrapper[4775]: E0321 04:49:38.660518 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:38 crc kubenswrapper[4775]: E0321 04:49:38.754951 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.900667 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.900958 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.900973 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.900989 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.900998 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:38Z","lastTransitionTime":"2026-03-21T04:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:38 crc kubenswrapper[4775]: E0321 04:49:38.916981 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:38Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.920501 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.920544 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.920558 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.920575 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.920586 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:38Z","lastTransitionTime":"2026-03-21T04:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:38 crc kubenswrapper[4775]: E0321 04:49:38.933577 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:38Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.936743 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.936772 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.936780 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.936792 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.936801 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:38Z","lastTransitionTime":"2026-03-21T04:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:38 crc kubenswrapper[4775]: E0321 04:49:38.947752 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:38Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.951051 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.951150 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.951165 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.951180 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.951192 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:38Z","lastTransitionTime":"2026-03-21T04:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:38 crc kubenswrapper[4775]: E0321 04:49:38.962778 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:38Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.965209 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.965236 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.965249 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.965264 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:38 crc kubenswrapper[4775]: I0321 04:49:38.965274 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:38Z","lastTransitionTime":"2026-03-21T04:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:38 crc kubenswrapper[4775]: E0321 04:49:38.975496 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:38Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:38 crc kubenswrapper[4775]: E0321 04:49:38.975609 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 21 04:49:39 crc kubenswrapper[4775]: I0321 04:49:39.660689 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:39 crc kubenswrapper[4775]: E0321 04:49:39.660950 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:39 crc kubenswrapper[4775]: I0321 04:49:39.661374 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:39 crc kubenswrapper[4775]: E0321 04:49:39.661473 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:39 crc kubenswrapper[4775]: I0321 04:49:39.661708 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:39 crc kubenswrapper[4775]: E0321 04:49:39.661783 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:40 crc kubenswrapper[4775]: I0321 04:49:40.661213 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:40 crc kubenswrapper[4775]: E0321 04:49:40.661332 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:41 crc kubenswrapper[4775]: I0321 04:49:41.660821 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:41 crc kubenswrapper[4775]: I0321 04:49:41.660952 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:41 crc kubenswrapper[4775]: E0321 04:49:41.661006 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:41 crc kubenswrapper[4775]: I0321 04:49:41.661088 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:41 crc kubenswrapper[4775]: E0321 04:49:41.661273 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:41 crc kubenswrapper[4775]: E0321 04:49:41.661338 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:42 crc kubenswrapper[4775]: I0321 04:49:42.660627 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:42 crc kubenswrapper[4775]: E0321 04:49:42.660808 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:43 crc kubenswrapper[4775]: I0321 04:49:43.660389 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:43 crc kubenswrapper[4775]: I0321 04:49:43.660382 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:43 crc kubenswrapper[4775]: E0321 04:49:43.660841 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:43 crc kubenswrapper[4775]: I0321 04:49:43.660521 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:43 crc kubenswrapper[4775]: E0321 04:49:43.660947 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:43 crc kubenswrapper[4775]: E0321 04:49:43.661186 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:43 crc kubenswrapper[4775]: I0321 04:49:43.673036 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 21 04:49:43 crc kubenswrapper[4775]: E0321 04:49:43.755746 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 04:49:44 crc kubenswrapper[4775]: I0321 04:49:44.660959 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:44 crc kubenswrapper[4775]: E0321 04:49:44.661084 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:45 crc kubenswrapper[4775]: I0321 04:49:45.660434 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:45 crc kubenswrapper[4775]: I0321 04:49:45.660527 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:45 crc kubenswrapper[4775]: I0321 04:49:45.660457 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:45 crc kubenswrapper[4775]: E0321 04:49:45.660593 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:45 crc kubenswrapper[4775]: E0321 04:49:45.660665 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:45 crc kubenswrapper[4775]: E0321 04:49:45.660761 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:46 crc kubenswrapper[4775]: I0321 04:49:46.660426 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:46 crc kubenswrapper[4775]: E0321 04:49:46.660558 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.660468 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.660517 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.660555 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:47 crc kubenswrapper[4775]: E0321 04:49:47.660652 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:47 crc kubenswrapper[4775]: E0321 04:49:47.660720 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:47 crc kubenswrapper[4775]: E0321 04:49:47.660771 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.672092 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:47Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.688314 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:47Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.699697 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:47Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.709198 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:47Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.721387 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:47Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.734509 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:47Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.747391 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:47Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.758427 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:47Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.771434 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:47Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.781627 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:47Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.795612 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:47Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.809187 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:47Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.822553 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:47Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.839476 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:34Z\\\",\\\"message\\\":\\\"eighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.21:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {58a148b3-0a7b-4412-b447-f87788c4883f}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0321 04:49:33.949601 7032 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-etcd-operator/metrics]} name:Service_openshift-etcd-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.188:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {53c717ca-2174-4315-bb03-c937a9c0d9b6}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0321 04:49:33.949663 7032 factory.go:656] Stopping watch factory\\\\nI0321 04:49:33.949677 7032 ovnkube.go:599] Stopped ovnkube\\\\nI0321 04:49:33.949692 7032 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0321 04:49:33.949770 7032 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mzqtk_openshift-ovn-kubernetes(a69d31f5-deeb-4860-be96-ed5547831685)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:47Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.850380 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a197f323-38cd-41a6-ae73-5d16706c9224\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f04df41528bb781c23382f46200bd21b1d0e6064f535d4d47cde4cabedef8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71042d514c08b225f10915e55a0ef253281a017b05412ca810a5fbbfee272bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a3e2381fa82bd51ba8be325e0a83a387ebd790d6a91283d604ab46f9ae7d45e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99787fe5e44de1b5adaa5d1b49dcb6db27a4176d232dc67119b5f633bfa4c948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99787fe5e44de1b5adaa5d1b49dcb6db27a4176d232dc67119b5f633bfa4c948\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:47Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.870915 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:47Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.881396 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:47Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:47 crc kubenswrapper[4775]: I0321 04:49:47.918968 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs\") pod \"network-metrics-daemon-xk9f5\" (UID: \"6920413a-2c51-466d-a16e-d14489ae0c6c\") " pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:47 crc kubenswrapper[4775]: E0321 04:49:47.919108 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 04:49:47 crc kubenswrapper[4775]: E0321 04:49:47.919204 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs podName:6920413a-2c51-466d-a16e-d14489ae0c6c nodeName:}" failed. No retries permitted until 2026-03-21 04:50:19.919184945 +0000 UTC m=+172.895648569 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs") pod "network-metrics-daemon-xk9f5" (UID: "6920413a-2c51-466d-a16e-d14489ae0c6c") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 04:49:48 crc kubenswrapper[4775]: I0321 04:49:48.661222 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:48 crc kubenswrapper[4775]: E0321 04:49:48.661390 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:48 crc kubenswrapper[4775]: E0321 04:49:48.757348 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.068188 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.068444 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.068623 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.068782 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.068923 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:49Z","lastTransitionTime":"2026-03-21T04:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:49 crc kubenswrapper[4775]: E0321 04:49:49.080636 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:49Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.083977 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.084003 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.084012 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.084024 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.084033 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:49Z","lastTransitionTime":"2026-03-21T04:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:49 crc kubenswrapper[4775]: E0321 04:49:49.094497 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:49Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.097982 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.098024 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.098059 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.098079 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.098093 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:49Z","lastTransitionTime":"2026-03-21T04:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:49 crc kubenswrapper[4775]: E0321 04:49:49.111383 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:49Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.115304 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.115345 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.115360 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.115431 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.115477 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:49Z","lastTransitionTime":"2026-03-21T04:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:49 crc kubenswrapper[4775]: E0321 04:49:49.130430 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:49Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.134339 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.134382 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.134394 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.134411 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.134426 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:49Z","lastTransitionTime":"2026-03-21T04:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:49 crc kubenswrapper[4775]: E0321 04:49:49.147530 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:49Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:49 crc kubenswrapper[4775]: E0321 04:49:49.147797 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.660972 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.660995 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.661398 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:49 crc kubenswrapper[4775]: I0321 04:49:49.661856 4775 scope.go:117] "RemoveContainer" containerID="a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b" Mar 21 04:49:49 crc kubenswrapper[4775]: E0321 04:49:49.661536 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:49 crc kubenswrapper[4775]: E0321 04:49:49.661879 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:49 crc kubenswrapper[4775]: E0321 04:49:49.661650 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:49 crc kubenswrapper[4775]: E0321 04:49:49.662085 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mzqtk_openshift-ovn-kubernetes(a69d31f5-deeb-4860-be96-ed5547831685)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" podUID="a69d31f5-deeb-4860-be96-ed5547831685" Mar 21 04:49:50 crc kubenswrapper[4775]: I0321 04:49:50.661448 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:50 crc kubenswrapper[4775]: E0321 04:49:50.661661 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:51 crc kubenswrapper[4775]: I0321 04:49:51.660811 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:51 crc kubenswrapper[4775]: E0321 04:49:51.661288 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:51 crc kubenswrapper[4775]: I0321 04:49:51.660897 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:51 crc kubenswrapper[4775]: E0321 04:49:51.661524 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:51 crc kubenswrapper[4775]: I0321 04:49:51.660813 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:51 crc kubenswrapper[4775]: E0321 04:49:51.662001 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:52 crc kubenswrapper[4775]: I0321 04:49:52.660803 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:52 crc kubenswrapper[4775]: E0321 04:49:52.661061 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:52 crc kubenswrapper[4775]: I0321 04:49:52.954957 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-556rg_e77ec218-42da-4f07-b214-184c4f3b20f3/kube-multus/0.log" Mar 21 04:49:52 crc kubenswrapper[4775]: I0321 04:49:52.955000 4775 generic.go:334] "Generic (PLEG): container finished" podID="e77ec218-42da-4f07-b214-184c4f3b20f3" containerID="77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf" exitCode=1 Mar 21 04:49:52 crc kubenswrapper[4775]: I0321 04:49:52.955027 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-556rg" event={"ID":"e77ec218-42da-4f07-b214-184c4f3b20f3","Type":"ContainerDied","Data":"77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf"} Mar 21 04:49:52 crc kubenswrapper[4775]: I0321 04:49:52.955393 4775 scope.go:117] "RemoveContainer" containerID="77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf" Mar 21 04:49:52 crc kubenswrapper[4775]: I0321 04:49:52.970831 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:52Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:52 crc kubenswrapper[4775]: I0321 04:49:52.991357 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:52Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.007215 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:53Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.021348 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:52Z\\\",\\\"message\\\":\\\"2026-03-21T04:49:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71b8a2db-de4c-4816-881e-455a501c594f\\\\n2026-03-21T04:49:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71b8a2db-de4c-4816-881e-455a501c594f to /host/opt/cni/bin/\\\\n2026-03-21T04:49:07Z [verbose] multus-daemon started\\\\n2026-03-21T04:49:07Z [verbose] Readiness Indicator file check\\\\n2026-03-21T04:49:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:53Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.031403 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:53Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.040408 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:53Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.052042 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:53Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.070377 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:53Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.082992 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:53Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.100981 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:53Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.114411 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:53Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.126318 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:53Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.137157 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:53Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.149036 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a197f323-38cd-41a6-ae73-5d16706c9224\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f04df41528bb781c23382f46200bd21b1d0e6064f535d4d47cde4cabedef8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71042d514c08b225f10915e55a0ef253281a017b05412ca810a5fbbfee272bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a3e2381fa82bd51ba8be325e0a83a387ebd790d6a91283d604ab46f9ae7d45e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99787fe5e44de1b5adaa5d1b49dcb6db27a4176d232dc67119b5f633bfa4c948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99787fe5e44de1b5adaa5d1b49dcb6db27a4176d232dc67119b5f633bfa4c948\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:53Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.168043 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:53Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.179969 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:53Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.198215 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:34Z\\\",\\\"message\\\":\\\"eighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.21:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {58a148b3-0a7b-4412-b447-f87788c4883f}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0321 04:49:33.949601 7032 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-etcd-operator/metrics]} name:Service_openshift-etcd-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.188:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {53c717ca-2174-4315-bb03-c937a9c0d9b6}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0321 04:49:33.949663 7032 factory.go:656] Stopping watch factory\\\\nI0321 04:49:33.949677 7032 ovnkube.go:599] Stopped ovnkube\\\\nI0321 04:49:33.949692 7032 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0321 04:49:33.949770 7032 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mzqtk_openshift-ovn-kubernetes(a69d31f5-deeb-4860-be96-ed5547831685)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:53Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.661035 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:53 crc kubenswrapper[4775]: E0321 04:49:53.661450 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.661063 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:53 crc kubenswrapper[4775]: E0321 04:49:53.661727 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.661038 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:53 crc kubenswrapper[4775]: E0321 04:49:53.661936 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:53 crc kubenswrapper[4775]: E0321 04:49:53.759554 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.961046 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-556rg_e77ec218-42da-4f07-b214-184c4f3b20f3/kube-multus/0.log" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.961184 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-556rg" event={"ID":"e77ec218-42da-4f07-b214-184c4f3b20f3","Type":"ContainerStarted","Data":"eaf83cd6f4a806265a871a5672f1a6939180308c2815b87ca254164f26d05aae"} Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.976869 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:53Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:53 crc kubenswrapper[4775]: I0321 04:49:53.994260 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:53Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:54 crc kubenswrapper[4775]: I0321 04:49:54.010260 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:54Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:54 crc kubenswrapper[4775]: I0321 04:49:54.029216 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:54Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:54 crc kubenswrapper[4775]: I0321 04:49:54.044110 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:54Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:54 crc kubenswrapper[4775]: I0321 04:49:54.059291 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:54Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:54 crc kubenswrapper[4775]: I0321 04:49:54.071909 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:54Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:54 crc kubenswrapper[4775]: I0321 04:49:54.084904 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a197f323-38cd-41a6-ae73-5d16706c9224\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f04df41528bb781c23382f46200bd21b1d0e6064f535d4d47cde4cabedef8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71042d514c08b225f10915e55a0ef253281a017b05412ca810a5fbbfee272bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a3e2381fa82bd51ba8be325e0a83a387ebd790d6a91283d604ab46f9ae7d45e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99787fe5e44de1b5adaa5d1b49dcb6db27a4176d232dc67119b5f633bfa4c948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99787fe5e44de1b5adaa5d1b49dcb6db27a4176d232dc67119b5f633bfa4c948\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:54Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:54 crc kubenswrapper[4775]: I0321 04:49:54.106304 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:54Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:54 crc kubenswrapper[4775]: I0321 04:49:54.120230 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:54Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:54 crc kubenswrapper[4775]: I0321 04:49:54.149275 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:34Z\\\",\\\"message\\\":\\\"eighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.21:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {58a148b3-0a7b-4412-b447-f87788c4883f}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0321 04:49:33.949601 7032 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-etcd-operator/metrics]} name:Service_openshift-etcd-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.188:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {53c717ca-2174-4315-bb03-c937a9c0d9b6}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0321 04:49:33.949663 7032 factory.go:656] Stopping watch factory\\\\nI0321 04:49:33.949677 7032 ovnkube.go:599] Stopped ovnkube\\\\nI0321 04:49:33.949692 7032 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0321 04:49:33.949770 7032 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mzqtk_openshift-ovn-kubernetes(a69d31f5-deeb-4860-be96-ed5547831685)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:54Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:54 crc kubenswrapper[4775]: I0321 04:49:54.165081 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:54Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:54 crc kubenswrapper[4775]: I0321 04:49:54.178260 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:54Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:54 crc kubenswrapper[4775]: I0321 04:49:54.192203 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:54Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:54 crc kubenswrapper[4775]: I0321 04:49:54.207853 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaf83cd6f4a806265a871a5672f1a6939180308c2815b87ca254164f26d05aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:52Z\\\",\\\"message\\\":\\\"2026-03-21T04:49:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71b8a2db-de4c-4816-881e-455a501c594f\\\\n2026-03-21T04:49:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71b8a2db-de4c-4816-881e-455a501c594f to /host/opt/cni/bin/\\\\n2026-03-21T04:49:07Z [verbose] multus-daemon started\\\\n2026-03-21T04:49:07Z [verbose] Readiness Indicator file check\\\\n2026-03-21T04:49:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:54Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:54 crc kubenswrapper[4775]: I0321 04:49:54.219268 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:54Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:54 crc kubenswrapper[4775]: I0321 04:49:54.232456 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:54Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:54 crc kubenswrapper[4775]: I0321 04:49:54.660806 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:54 crc kubenswrapper[4775]: E0321 04:49:54.661248 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:54 crc kubenswrapper[4775]: I0321 04:49:54.671828 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 21 04:49:55 crc kubenswrapper[4775]: I0321 04:49:55.661165 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:55 crc kubenswrapper[4775]: I0321 04:49:55.661191 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:55 crc kubenswrapper[4775]: I0321 04:49:55.661217 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:55 crc kubenswrapper[4775]: E0321 04:49:55.661306 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:55 crc kubenswrapper[4775]: E0321 04:49:55.661396 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:55 crc kubenswrapper[4775]: E0321 04:49:55.661482 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:56 crc kubenswrapper[4775]: I0321 04:49:56.415402 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:49:56 crc kubenswrapper[4775]: E0321 04:49:56.415527 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:00.415507516 +0000 UTC m=+213.391971140 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:49:56 crc kubenswrapper[4775]: I0321 04:49:56.415650 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:56 crc kubenswrapper[4775]: I0321 04:49:56.415760 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:56 crc kubenswrapper[4775]: E0321 04:49:56.415795 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 04:49:56 crc kubenswrapper[4775]: E0321 04:49:56.415855 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 04:51:00.415842895 +0000 UTC m=+213.392306519 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 04:49:56 crc kubenswrapper[4775]: E0321 04:49:56.415889 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 04:49:56 crc kubenswrapper[4775]: E0321 04:49:56.415951 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 04:51:00.415940298 +0000 UTC m=+213.392403922 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 04:49:56 crc kubenswrapper[4775]: I0321 04:49:56.516314 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:56 crc kubenswrapper[4775]: I0321 04:49:56.516357 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:56 crc kubenswrapper[4775]: E0321 04:49:56.516472 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 04:49:56 crc kubenswrapper[4775]: E0321 04:49:56.516489 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 04:49:56 crc kubenswrapper[4775]: E0321 04:49:56.516500 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:49:56 crc kubenswrapper[4775]: E0321 04:49:56.516557 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-21 04:51:00.516542081 +0000 UTC m=+213.493005705 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:49:56 crc kubenswrapper[4775]: E0321 04:49:56.516627 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 04:49:56 crc kubenswrapper[4775]: E0321 04:49:56.516692 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 04:49:56 crc kubenswrapper[4775]: E0321 04:49:56.516714 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:49:56 crc kubenswrapper[4775]: E0321 04:49:56.516801 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-21 04:51:00.516775128 +0000 UTC m=+213.493238752 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 04:49:56 crc kubenswrapper[4775]: I0321 04:49:56.660851 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:56 crc kubenswrapper[4775]: E0321 04:49:56.660996 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:56 crc kubenswrapper[4775]: I0321 04:49:56.669492 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.660743 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.660801 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.660957 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:57 crc kubenswrapper[4775]: E0321 04:49:57.660944 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:49:57 crc kubenswrapper[4775]: E0321 04:49:57.661155 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:57 crc kubenswrapper[4775]: E0321 04:49:57.661239 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.675393 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:57Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.694892 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:57Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.709256 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:57Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.720730 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:57Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.731967 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a197f323-38cd-41a6-ae73-5d16706c9224\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f04df41528bb781c23382f46200bd21b1d0e6064f535d4d47cde4cabedef8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71042d514c08b225f10915e55a0ef253281a017b05412ca810a5fbbfee272bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a3e2381fa82bd51ba8be325e0a83a387ebd790d6a91283d604ab46f9ae7d45e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99787fe5e44de1b5adaa5d1b49dcb6db27a4176d232dc67119b5f633bfa4c948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99787fe5e44de1b5adaa5d1b49dcb6db27a4176d232dc67119b5f633bfa4c948\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:57Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.756687 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:57Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.774328 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:57Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.795492 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:34Z\\\",\\\"message\\\":\\\"eighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.21:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {58a148b3-0a7b-4412-b447-f87788c4883f}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0321 04:49:33.949601 7032 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-etcd-operator/metrics]} name:Service_openshift-etcd-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.188:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {53c717ca-2174-4315-bb03-c937a9c0d9b6}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0321 04:49:33.949663 7032 factory.go:656] Stopping watch factory\\\\nI0321 04:49:33.949677 7032 ovnkube.go:599] Stopped ovnkube\\\\nI0321 04:49:33.949692 7032 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0321 04:49:33.949770 7032 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mzqtk_openshift-ovn-kubernetes(a69d31f5-deeb-4860-be96-ed5547831685)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:57Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.805249 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdcbc41e-a246-40f6-bf50-36d3c217315a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf88511498eb0f6182819ecb78616730fa9bfbe187057ea891f3ba73e550287\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375572b0f0f7a6917d62ef1138d8651eb2d6118d420fc647d1d24c29e793f17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://375572b0f0f7a6917d62ef1138d8651eb2d6118d420fc647d1d24c29e793f17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:57Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.816327 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:57Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.826398 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:57Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.835873 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:57Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.846206 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaf83cd6f4a806265a871a5672f1a6939180308c2815b87ca254164f26d05aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:52Z\\\",\\\"message\\\":\\\"2026-03-21T04:49:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71b8a2db-de4c-4816-881e-455a501c594f\\\\n2026-03-21T04:49:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71b8a2db-de4c-4816-881e-455a501c594f to /host/opt/cni/bin/\\\\n2026-03-21T04:49:07Z [verbose] multus-daemon started\\\\n2026-03-21T04:49:07Z [verbose] Readiness Indicator file check\\\\n2026-03-21T04:49:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:57Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.853963 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:57Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.862092 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:57Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.872963 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"058178a5-0360-40e4-9baf-5478bac349dd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5615480a5b452c6bce19fa04738bb5d68984f62258c9486fc2ae787cfb9b7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://851dab44c58218d203e180465f8a216bfeec8271e78102856d11545520feabb4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:47:55Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0321 04:47:29.750929 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0321 04:47:29.753390 1 observer_polling.go:159] Starting file observer\\\\nI0321 04:47:29.787802 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0321 04:47:29.793292 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0321 04:47:55.400421 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0321 04:47:55.400543 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75cf7f90f0971bce2e51690954080e297523b92a5588c8effa75832879f0dd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cb44c7daffbeddbb39314ec11bd3235996e19b520875337dcac1ffd14f35953\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9299f9db8cbdd2dbd04e47bd1eba4f95724ec3dd04ec5083564dc0c38a960457\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:57Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.885096 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:57Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.898951 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:57Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:57 crc kubenswrapper[4775]: I0321 04:49:57.910172 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:57Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:58 crc kubenswrapper[4775]: I0321 04:49:58.660591 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:49:58 crc kubenswrapper[4775]: E0321 04:49:58.661093 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:49:58 crc kubenswrapper[4775]: E0321 04:49:58.761704 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.501702 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.501758 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.501768 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.501786 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.501799 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:59Z","lastTransitionTime":"2026-03-21T04:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:59 crc kubenswrapper[4775]: E0321 04:49:59.516481 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:59Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.519929 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.519973 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.519981 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.519995 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.520004 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:59Z","lastTransitionTime":"2026-03-21T04:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:59 crc kubenswrapper[4775]: E0321 04:49:59.534289 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:59Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.538611 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.538649 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.538659 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.538675 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.538686 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:59Z","lastTransitionTime":"2026-03-21T04:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:59 crc kubenswrapper[4775]: E0321 04:49:59.551676 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:59Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.555283 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.555316 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.555326 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.555340 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.555358 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:59Z","lastTransitionTime":"2026-03-21T04:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:59 crc kubenswrapper[4775]: E0321 04:49:59.570165 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:59Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.573918 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.573972 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.573984 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.574000 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.574011 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:49:59Z","lastTransitionTime":"2026-03-21T04:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:49:59 crc kubenswrapper[4775]: E0321 04:49:59.586568 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:49:59Z is after 2025-08-24T17:21:41Z" Mar 21 04:49:59 crc kubenswrapper[4775]: E0321 04:49:59.586693 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.661427 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:49:59 crc kubenswrapper[4775]: E0321 04:49:59.661557 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.661880 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:49:59 crc kubenswrapper[4775]: E0321 04:49:59.661947 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:49:59 crc kubenswrapper[4775]: I0321 04:49:59.662111 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:49:59 crc kubenswrapper[4775]: E0321 04:49:59.662200 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:00 crc kubenswrapper[4775]: I0321 04:50:00.660456 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:00 crc kubenswrapper[4775]: E0321 04:50:00.660561 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:01 crc kubenswrapper[4775]: I0321 04:50:01.660675 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:01 crc kubenswrapper[4775]: E0321 04:50:01.660821 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:01 crc kubenswrapper[4775]: I0321 04:50:01.660900 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:01 crc kubenswrapper[4775]: I0321 04:50:01.661326 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:01 crc kubenswrapper[4775]: E0321 04:50:01.661270 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:01 crc kubenswrapper[4775]: E0321 04:50:01.661488 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:02 crc kubenswrapper[4775]: I0321 04:50:02.660962 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:02 crc kubenswrapper[4775]: E0321 04:50:02.661178 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:03 crc kubenswrapper[4775]: I0321 04:50:03.661016 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:03 crc kubenswrapper[4775]: I0321 04:50:03.661014 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:03 crc kubenswrapper[4775]: I0321 04:50:03.661376 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:03 crc kubenswrapper[4775]: E0321 04:50:03.661239 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:03 crc kubenswrapper[4775]: E0321 04:50:03.661477 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:03 crc kubenswrapper[4775]: E0321 04:50:03.661680 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:03 crc kubenswrapper[4775]: E0321 04:50:03.763619 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 04:50:04 crc kubenswrapper[4775]: I0321 04:50:04.660435 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:04 crc kubenswrapper[4775]: E0321 04:50:04.660616 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:04 crc kubenswrapper[4775]: I0321 04:50:04.661394 4775 scope.go:117] "RemoveContainer" containerID="a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b" Mar 21 04:50:04 crc kubenswrapper[4775]: I0321 04:50:04.996858 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzqtk_a69d31f5-deeb-4860-be96-ed5547831685/ovnkube-controller/2.log" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.000630 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerStarted","Data":"7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a"} Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.002590 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.020406 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.036050 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.050759 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.064304 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.076578 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a197f323-38cd-41a6-ae73-5d16706c9224\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f04df41528bb781c23382f46200bd21b1d0e6064f535d4d47cde4cabedef8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71042d514c08b225f10915e55a0ef253281a017b05412ca810a5fbbfee272bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a3e2381fa82bd51ba8be325e0a83a387ebd790d6a91283d604ab46f9ae7d45e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99787fe5e44de1b5adaa5d1b49dcb6db27a4176d232dc67119b5f633bfa4c948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99787fe5e44de1b5adaa5d1b49dcb6db27a4176d232dc67119b5f633bfa4c948\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.099401 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.111714 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.129141 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:34Z\\\",\\\"message\\\":\\\"eighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.21:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {58a148b3-0a7b-4412-b447-f87788c4883f}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0321 04:49:33.949601 7032 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-etcd-operator/metrics]} name:Service_openshift-etcd-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.188:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {53c717ca-2174-4315-bb03-c937a9c0d9b6}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0321 04:49:33.949663 7032 factory.go:656] Stopping watch factory\\\\nI0321 04:49:33.949677 7032 ovnkube.go:599] Stopped ovnkube\\\\nI0321 04:49:33.949692 7032 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0321 04:49:33.949770 7032 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.139983 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.153384 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdcbc41e-a246-40f6-bf50-36d3c217315a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf88511498eb0f6182819ecb78616730fa9bfbe187057ea891f3ba73e550287\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375572b0f0f7a6917d62ef1138d8651eb2d6118d420fc647d1d24c29e793f17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://375572b0f0f7a6917d62ef1138d8651eb2d6118d420fc647d1d24c29e793f17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.167381 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.193137 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.204345 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.242680 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaf83cd6f4a806265a871a5672f1a6939180308c2815b87ca254164f26d05aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:52Z\\\",\\\"message\\\":\\\"2026-03-21T04:49:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71b8a2db-de4c-4816-881e-455a501c594f\\\\n2026-03-21T04:49:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71b8a2db-de4c-4816-881e-455a501c594f to /host/opt/cni/bin/\\\\n2026-03-21T04:49:07Z [verbose] multus-daemon started\\\\n2026-03-21T04:49:07Z [verbose] Readiness Indicator file check\\\\n2026-03-21T04:49:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.265796 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.287917 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"058178a5-0360-40e4-9baf-5478bac349dd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5615480a5b452c6bce19fa04738bb5d68984f62258c9486fc2ae787cfb9b7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://851dab44c58218d203e180465f8a216bfeec8271e78102856d11545520feabb4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:47:55Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0321 04:47:29.750929 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0321 04:47:29.753390 1 observer_polling.go:159] Starting file observer\\\\nI0321 04:47:29.787802 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0321 04:47:29.793292 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0321 04:47:55.400421 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0321 04:47:55.400543 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75cf7f90f0971bce2e51690954080e297523b92a5588c8effa75832879f0dd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cb44c7daffbeddbb39314ec11bd3235996e19b520875337dcac1ffd14f35953\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9299f9db8cbdd2dbd04e47bd1eba4f95724ec3dd04ec5083564dc0c38a960457\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.301264 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.316248 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.327322 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.660883 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.660919 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:05 crc kubenswrapper[4775]: I0321 04:50:05.661047 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:05 crc kubenswrapper[4775]: E0321 04:50:05.661283 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:05 crc kubenswrapper[4775]: E0321 04:50:05.661408 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:05 crc kubenswrapper[4775]: E0321 04:50:05.661531 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.006036 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzqtk_a69d31f5-deeb-4860-be96-ed5547831685/ovnkube-controller/3.log" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.006751 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzqtk_a69d31f5-deeb-4860-be96-ed5547831685/ovnkube-controller/2.log" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.009939 4775 generic.go:334] "Generic (PLEG): container finished" podID="a69d31f5-deeb-4860-be96-ed5547831685" containerID="7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a" exitCode=1 Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.009983 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerDied","Data":"7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a"} Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.010030 4775 scope.go:117] "RemoveContainer" containerID="a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.010687 4775 scope.go:117] "RemoveContainer" containerID="7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a" Mar 21 04:50:06 crc kubenswrapper[4775]: E0321 04:50:06.010951 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mzqtk_openshift-ovn-kubernetes(a69d31f5-deeb-4860-be96-ed5547831685)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" podUID="a69d31f5-deeb-4860-be96-ed5547831685" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.031664 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.047282 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.060317 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.073089 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.095234 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.108641 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.128349 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a8e35d65afe6149f7a4d22b0c351839afb61b412d54dfdcd93311c341d263b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:34Z\\\",\\\"message\\\":\\\"eighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.21:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {58a148b3-0a7b-4412-b447-f87788c4883f}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0321 04:49:33.949601 7032 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-etcd-operator/metrics]} name:Service_openshift-etcd-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.188:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {53c717ca-2174-4315-bb03-c937a9c0d9b6}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0321 04:49:33.949663 7032 factory.go:656] Stopping watch factory\\\\nI0321 04:49:33.949677 7032 ovnkube.go:599] Stopped ovnkube\\\\nI0321 04:49:33.949692 7032 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0321 04:49:33.949770 7032 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:50:05Z\\\",\\\"message\\\":\\\"05.638184 7377 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nF0321 04:50:05.638209 7377 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z]\\\\nI0321 04:50:05.638213 7377 obj_retry.go:409] Going to retry *v1.Pod resource setup for 16 objects: [openshift-kube-apiserver/kube-apiserver-crc openshift-machine-config-operator/kube-rbac-proxy-crio-crc openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz openshift-multus/multus-556rg openshift-network-node-identity/network-node-identity-vrzqb openshift-network-opera\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.140190 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a197f323-38cd-41a6-ae73-5d16706c9224\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f04df41528bb781c23382f46200bd21b1d0e6064f535d4d47cde4cabedef8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71042d514c08b225f10915e55a0ef253281a017b05412ca810a5fbbfee272bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a3e2381fa82bd51ba8be325e0a83a387ebd790d6a91283d604ab46f9ae7d45e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99787fe5e44de1b5adaa5d1b49dcb6db27a4176d232dc67119b5f633bfa4c948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99787fe5e44de1b5adaa5d1b49dcb6db27a4176d232dc67119b5f633bfa4c948\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.154517 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.168105 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.178386 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.190429 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaf83cd6f4a806265a871a5672f1a6939180308c2815b87ca254164f26d05aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:52Z\\\",\\\"message\\\":\\\"2026-03-21T04:49:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71b8a2db-de4c-4816-881e-455a501c594f\\\\n2026-03-21T04:49:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71b8a2db-de4c-4816-881e-455a501c594f to /host/opt/cni/bin/\\\\n2026-03-21T04:49:07Z [verbose] multus-daemon started\\\\n2026-03-21T04:49:07Z [verbose] Readiness Indicator file check\\\\n2026-03-21T04:49:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.200295 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.211633 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.222209 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdcbc41e-a246-40f6-bf50-36d3c217315a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf88511498eb0f6182819ecb78616730fa9bfbe187057ea891f3ba73e550287\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375572b0f0f7a6917d62ef1138d8651eb2d6118d420fc647d1d24c29e793f17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://375572b0f0f7a6917d62ef1138d8651eb2d6118d420fc647d1d24c29e793f17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.232947 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.247817 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.260446 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.273451 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"058178a5-0360-40e4-9baf-5478bac349dd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5615480a5b452c6bce19fa04738bb5d68984f62258c9486fc2ae787cfb9b7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://851dab44c58218d203e180465f8a216bfeec8271e78102856d11545520feabb4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:47:55Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0321 04:47:29.750929 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0321 04:47:29.753390 1 observer_polling.go:159] Starting file observer\\\\nI0321 04:47:29.787802 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0321 04:47:29.793292 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0321 04:47:55.400421 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0321 04:47:55.400543 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75cf7f90f0971bce2e51690954080e297523b92a5588c8effa75832879f0dd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cb44c7daffbeddbb39314ec11bd3235996e19b520875337dcac1ffd14f35953\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9299f9db8cbdd2dbd04e47bd1eba4f95724ec3dd04ec5083564dc0c38a960457\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:06Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:06 crc kubenswrapper[4775]: I0321 04:50:06.660974 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:06 crc kubenswrapper[4775]: E0321 04:50:06.661215 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.014309 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzqtk_a69d31f5-deeb-4860-be96-ed5547831685/ovnkube-controller/3.log" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.018186 4775 scope.go:117] "RemoveContainer" containerID="7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a" Mar 21 04:50:07 crc kubenswrapper[4775]: E0321 04:50:07.018355 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mzqtk_openshift-ovn-kubernetes(a69d31f5-deeb-4860-be96-ed5547831685)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" podUID="a69d31f5-deeb-4860-be96-ed5547831685" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.034080 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.046916 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.057152 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.072836 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.093730 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.105345 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.132272 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:50:05Z\\\",\\\"message\\\":\\\"05.638184 7377 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nF0321 04:50:05.638209 7377 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z]\\\\nI0321 04:50:05.638213 7377 obj_retry.go:409] Going to retry *v1.Pod resource setup for 16 objects: [openshift-kube-apiserver/kube-apiserver-crc openshift-machine-config-operator/kube-rbac-proxy-crio-crc openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz openshift-multus/multus-556rg openshift-network-node-identity/network-node-identity-vrzqb openshift-network-opera\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:50:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mzqtk_openshift-ovn-kubernetes(a69d31f5-deeb-4860-be96-ed5547831685)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.145809 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a197f323-38cd-41a6-ae73-5d16706c9224\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f04df41528bb781c23382f46200bd21b1d0e6064f535d4d47cde4cabedef8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71042d514c08b225f10915e55a0ef253281a017b05412ca810a5fbbfee272bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a3e2381fa82bd51ba8be325e0a83a387ebd790d6a91283d604ab46f9ae7d45e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99787fe5e44de1b5adaa5d1b49dcb6db27a4176d232dc67119b5f633bfa4c948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99787fe5e44de1b5adaa5d1b49dcb6db27a4176d232dc67119b5f633bfa4c948\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.160567 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.173484 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.186355 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.198608 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaf83cd6f4a806265a871a5672f1a6939180308c2815b87ca254164f26d05aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:52Z\\\",\\\"message\\\":\\\"2026-03-21T04:49:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71b8a2db-de4c-4816-881e-455a501c594f\\\\n2026-03-21T04:49:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71b8a2db-de4c-4816-881e-455a501c594f to /host/opt/cni/bin/\\\\n2026-03-21T04:49:07Z [verbose] multus-daemon started\\\\n2026-03-21T04:49:07Z [verbose] Readiness Indicator file check\\\\n2026-03-21T04:49:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.209213 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.219945 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.229670 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdcbc41e-a246-40f6-bf50-36d3c217315a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf88511498eb0f6182819ecb78616730fa9bfbe187057ea891f3ba73e550287\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375572b0f0f7a6917d62ef1138d8651eb2d6118d420fc647d1d24c29e793f17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://375572b0f0f7a6917d62ef1138d8651eb2d6118d420fc647d1d24c29e793f17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.239277 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.254554 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.266547 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.277471 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"058178a5-0360-40e4-9baf-5478bac349dd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5615480a5b452c6bce19fa04738bb5d68984f62258c9486fc2ae787cfb9b7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://851dab44c58218d203e180465f8a216bfeec8271e78102856d11545520feabb4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:47:55Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0321 04:47:29.750929 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0321 04:47:29.753390 1 observer_polling.go:159] Starting file observer\\\\nI0321 04:47:29.787802 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0321 04:47:29.793292 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0321 04:47:55.400421 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0321 04:47:55.400543 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75cf7f90f0971bce2e51690954080e297523b92a5588c8effa75832879f0dd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cb44c7daffbeddbb39314ec11bd3235996e19b520875337dcac1ffd14f35953\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9299f9db8cbdd2dbd04e47bd1eba4f95724ec3dd04ec5083564dc0c38a960457\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.661327 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.661375 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:07 crc kubenswrapper[4775]: E0321 04:50:07.661475 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.661579 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:07 crc kubenswrapper[4775]: E0321 04:50:07.661655 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:07 crc kubenswrapper[4775]: E0321 04:50:07.661807 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.686183 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.705673 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.722638 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.738249 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.752294 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a197f323-38cd-41a6-ae73-5d16706c9224\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f04df41528bb781c23382f46200bd21b1d0e6064f535d4d47cde4cabedef8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71042d514c08b225f10915e55a0ef253281a017b05412ca810a5fbbfee272bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a3e2381fa82bd51ba8be325e0a83a387ebd790d6a91283d604ab46f9ae7d45e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99787fe5e44de1b5adaa5d1b49dcb6db27a4176d232dc67119b5f633bfa4c948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99787fe5e44de1b5adaa5d1b49dcb6db27a4176d232dc67119b5f633bfa4c948\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.779550 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.795901 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.816571 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:50:05Z\\\",\\\"message\\\":\\\"05.638184 7377 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nF0321 04:50:05.638209 7377 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z]\\\\nI0321 04:50:05.638213 7377 obj_retry.go:409] Going to retry *v1.Pod resource setup for 16 objects: [openshift-kube-apiserver/kube-apiserver-crc openshift-machine-config-operator/kube-rbac-proxy-crio-crc openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz openshift-multus/multus-556rg openshift-network-node-identity/network-node-identity-vrzqb openshift-network-opera\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:50:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mzqtk_openshift-ovn-kubernetes(a69d31f5-deeb-4860-be96-ed5547831685)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.831650 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.846190 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdcbc41e-a246-40f6-bf50-36d3c217315a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf88511498eb0f6182819ecb78616730fa9bfbe187057ea891f3ba73e550287\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375572b0f0f7a6917d62ef1138d8651eb2d6118d420fc647d1d24c29e793f17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://375572b0f0f7a6917d62ef1138d8651eb2d6118d420fc647d1d24c29e793f17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.861922 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.876246 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.890930 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.907295 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaf83cd6f4a806265a871a5672f1a6939180308c2815b87ca254164f26d05aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:52Z\\\",\\\"message\\\":\\\"2026-03-21T04:49:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71b8a2db-de4c-4816-881e-455a501c594f\\\\n2026-03-21T04:49:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71b8a2db-de4c-4816-881e-455a501c594f to /host/opt/cni/bin/\\\\n2026-03-21T04:49:07Z [verbose] multus-daemon started\\\\n2026-03-21T04:49:07Z [verbose] Readiness Indicator file check\\\\n2026-03-21T04:49:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.923974 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.942949 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"058178a5-0360-40e4-9baf-5478bac349dd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5615480a5b452c6bce19fa04738bb5d68984f62258c9486fc2ae787cfb9b7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://851dab44c58218d203e180465f8a216bfeec8271e78102856d11545520feabb4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:47:55Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0321 04:47:29.750929 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0321 04:47:29.753390 1 observer_polling.go:159] Starting file observer\\\\nI0321 04:47:29.787802 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0321 04:47:29.793292 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0321 04:47:55.400421 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0321 04:47:55.400543 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75cf7f90f0971bce2e51690954080e297523b92a5588c8effa75832879f0dd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cb44c7daffbeddbb39314ec11bd3235996e19b520875337dcac1ffd14f35953\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9299f9db8cbdd2dbd04e47bd1eba4f95724ec3dd04ec5083564dc0c38a960457\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.955754 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.972895 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:07 crc kubenswrapper[4775]: I0321 04:50:07.989021 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:07Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:08 crc kubenswrapper[4775]: I0321 04:50:08.660363 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:08 crc kubenswrapper[4775]: E0321 04:50:08.660693 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:08 crc kubenswrapper[4775]: E0321 04:50:08.764453 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.661203 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.661203 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:09 crc kubenswrapper[4775]: E0321 04:50:09.661367 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:09 crc kubenswrapper[4775]: E0321 04:50:09.661465 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.661222 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:09 crc kubenswrapper[4775]: E0321 04:50:09.661568 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.823469 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.823509 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.823521 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.823536 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.823546 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:50:09Z","lastTransitionTime":"2026-03-21T04:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:50:09 crc kubenswrapper[4775]: E0321 04:50:09.836071 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.840111 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.840161 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.840174 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.840188 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.840198 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:50:09Z","lastTransitionTime":"2026-03-21T04:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:50:09 crc kubenswrapper[4775]: E0321 04:50:09.854182 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.858390 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.858423 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.858643 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.858663 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.858675 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:50:09Z","lastTransitionTime":"2026-03-21T04:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:50:09 crc kubenswrapper[4775]: E0321 04:50:09.874367 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.877515 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.877545 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.877558 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.877574 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.877586 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:50:09Z","lastTransitionTime":"2026-03-21T04:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:50:09 crc kubenswrapper[4775]: E0321 04:50:09.888281 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.891372 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.891408 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.891420 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.891436 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:50:09 crc kubenswrapper[4775]: I0321 04:50:09.891448 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:50:09Z","lastTransitionTime":"2026-03-21T04:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:50:09 crc kubenswrapper[4775]: E0321 04:50:09.902985 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:09Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:09 crc kubenswrapper[4775]: E0321 04:50:09.903107 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 21 04:50:10 crc kubenswrapper[4775]: I0321 04:50:10.660364 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:10 crc kubenswrapper[4775]: E0321 04:50:10.660761 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:11 crc kubenswrapper[4775]: I0321 04:50:11.661008 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:11 crc kubenswrapper[4775]: I0321 04:50:11.661043 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:11 crc kubenswrapper[4775]: I0321 04:50:11.661105 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:11 crc kubenswrapper[4775]: E0321 04:50:11.661246 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:11 crc kubenswrapper[4775]: E0321 04:50:11.661331 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:11 crc kubenswrapper[4775]: E0321 04:50:11.661408 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:12 crc kubenswrapper[4775]: I0321 04:50:12.660392 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:12 crc kubenswrapper[4775]: E0321 04:50:12.660581 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:13 crc kubenswrapper[4775]: I0321 04:50:13.660462 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:13 crc kubenswrapper[4775]: E0321 04:50:13.660618 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:13 crc kubenswrapper[4775]: I0321 04:50:13.660728 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:13 crc kubenswrapper[4775]: E0321 04:50:13.660799 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:13 crc kubenswrapper[4775]: I0321 04:50:13.660875 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:13 crc kubenswrapper[4775]: E0321 04:50:13.660932 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:13 crc kubenswrapper[4775]: E0321 04:50:13.766214 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 04:50:14 crc kubenswrapper[4775]: I0321 04:50:14.660851 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:14 crc kubenswrapper[4775]: E0321 04:50:14.661007 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:15 crc kubenswrapper[4775]: I0321 04:50:15.661243 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:15 crc kubenswrapper[4775]: I0321 04:50:15.661375 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:15 crc kubenswrapper[4775]: I0321 04:50:15.661255 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:15 crc kubenswrapper[4775]: E0321 04:50:15.661368 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:15 crc kubenswrapper[4775]: E0321 04:50:15.661549 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:15 crc kubenswrapper[4775]: E0321 04:50:15.661576 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:16 crc kubenswrapper[4775]: I0321 04:50:16.660734 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:16 crc kubenswrapper[4775]: E0321 04:50:16.661078 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.660933 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.661005 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.660939 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:17 crc kubenswrapper[4775]: E0321 04:50:17.661204 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:17 crc kubenswrapper[4775]: E0321 04:50:17.661326 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:17 crc kubenswrapper[4775]: E0321 04:50:17.661431 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.685785 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a84635b7-13d6-414b-b564-df905a75f78e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd489dc7122bc77a71fedd534450c46432fd565c47c5a1bf0adbdbf692ff5846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b2cb55b98ba78b43ee20318d2240093a0ac190d08959d632fae08c0503d987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef017956ac1ac2629ec584876a1730d53fdb87672733c03f24391ba91530939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://202a0f82cbe6c8445fcb51a0878e276075a419dec4a93210947380a97da1bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf1328894b894eaf46c2d09d835b36aa7e931a51a0ab34bc0c7bb9b7b611e9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1474ef03c902f50e4c4565e4e251b19a9126b43e3918cc89871b86a499fa057\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://027439ea9557ee91465a3e603ef3d3d050016c9a92b039ede3cf6c6c17813539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336231c7d1d5fb22fd7ce69748f0263b9b20a45a6c60d3db6864d29960909645\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.700526 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.723737 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a69d31f5-deeb-4860-be96-ed5547831685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:50:05Z\\\",\\\"message\\\":\\\"05.638184 7377 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nF0321 04:50:05.638209 7377 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:05Z is after 2025-08-24T17:21:41Z]\\\\nI0321 04:50:05.638213 7377 obj_retry.go:409] Going to retry *v1.Pod resource setup for 16 objects: [openshift-kube-apiserver/kube-apiserver-crc openshift-machine-config-operator/kube-rbac-proxy-crio-crc openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz openshift-multus/multus-556rg openshift-network-node-identity/network-node-identity-vrzqb openshift-network-opera\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:50:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mzqtk_openshift-ovn-kubernetes(a69d31f5-deeb-4860-be96-ed5547831685)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7w6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzqtk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.738225 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a197f323-38cd-41a6-ae73-5d16706c9224\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f04df41528bb781c23382f46200bd21b1d0e6064f535d4d47cde4cabedef8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71042d514c08b225f10915e55a0ef253281a017b05412ca810a5fbbfee272bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a3e2381fa82bd51ba8be325e0a83a387ebd790d6a91283d604ab46f9ae7d45e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99787fe5e44de1b5adaa5d1b49dcb6db27a4176d232dc67119b5f633bfa4c948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99787fe5e44de1b5adaa5d1b49dcb6db27a4176d232dc67119b5f633bfa4c948\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.755875 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e4957b1b3b81f43f30ab64a060d6535b875e1aa8a77710f7c258245be089b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.771664 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ead4a6d7f69da8214783f61c88eef102aaa4b3cc06971e669dda923139ecbd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e479f2f8f2d9e9576d61a6e3ab302409f350cc28c555046c1ca6431307afaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.788541 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.805264 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-556rg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e77ec218-42da-4f07-b214-184c4f3b20f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaf83cd6f4a806265a871a5672f1a6939180308c2815b87ca254164f26d05aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-21T04:49:52Z\\\",\\\"message\\\":\\\"2026-03-21T04:49:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71b8a2db-de4c-4816-881e-455a501c594f\\\\n2026-03-21T04:49:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71b8a2db-de4c-4816-881e-455a501c594f to /host/opt/cni/bin/\\\\n2026-03-21T04:49:07Z [verbose] multus-daemon started\\\\n2026-03-21T04:49:07Z [verbose] Readiness Indicator file check\\\\n2026-03-21T04:49:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vft22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-556rg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.817927 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kr988" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc87971f-e8fc-454d-8513-957a0bbad389\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88f7eb6f5b1fecd00254bb66366d1fa1d90231b1927926acd0cc69bbb2f45fc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2kj8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kr988\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.829946 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6920413a-2c51-466d-a16e-d14489ae0c6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h68dp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xk9f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.843266 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdcbc41e-a246-40f6-bf50-36d3c217315a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf88511498eb0f6182819ecb78616730fa9bfbe187057ea891f3ba73e550287\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://375572b0f0f7a6917d62ef1138d8651eb2d6118d420fc647d1d24c29e793f17d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://375572b0f0f7a6917d62ef1138d8651eb2d6118d420fc647d1d24c29e793f17d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.857005 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-khh7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70cca5ae-9de4-4933-a6f8-4a23ab711bbf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb124fa0b2fc2582e8badf94a43d71a33b8865c5c466ab3bdcbb57e696de627e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bwtgh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-khh7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.873579 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kldzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"957cde70-ca20-438a-a4bf-42481dddb2db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12fc5d649ea6218fcab145822c0f2cd86e0513c53c81213eb7ed68fb3afde674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f60b38b40f94cc0e8e5cb4be791449552e2a54d23904a16227672eed7b3b18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fb5be104ab810117cde44c92e3a7dfe6ea7d7d1c728d89b6172379ecb1bbc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7261c2582b4756b7bf4c3acf67391dd7f0c478ed849b253ba9c511876a00785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://362f187c90be3c63ca293d844a192b1ddcdbca38ac34a0f59ce1faeb048e0de5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56b8be7f3b8b80c96c9bc89fcf927dbfefabd735ed217d85ac200b42b8d069fd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b07461492cde5a45e3db3c269dc79fb45d9340b15b69375d11512fba609750f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqnl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kldzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.886281 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38cb395c-744c-4c2e-9e32-b6cb206a9c5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99822baef16853fef450fdf2be6c8bd67171afc4b0a34e172b108f57738717d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226b410d4d80ed4dad83cf4c8e3d5c470f8d2cbae7852a11fde40d35d00ee235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4d4t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-62jtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.898754 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"058178a5-0360-40e4-9baf-5478bac349dd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5615480a5b452c6bce19fa04738bb5d68984f62258c9486fc2ae787cfb9b7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://851dab44c58218d203e180465f8a216bfeec8271e78102856d11545520feabb4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:47:55Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0321 04:47:29.750929 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0321 04:47:29.753390 1 observer_polling.go:159] Starting file observer\\\\nI0321 04:47:29.787802 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0321 04:47:29.793292 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0321 04:47:55.400421 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0321 04:47:55.400543 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75cf7f90f0971bce2e51690954080e297523b92a5588c8effa75832879f0dd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cb44c7daffbeddbb39314ec11bd3235996e19b520875337dcac1ffd14f35953\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9299f9db8cbdd2dbd04e47bd1eba4f95724ec3dd04ec5083564dc0c38a960457\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.912605 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.925631 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:48:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb93fe033377ef7e4bee2751d04558c879dea241df496c9222e2574272922143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.938025 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cffcf487-ef41-4395-81eb-e5e6358f4a32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3317fa0c2a16ea0f8c6be9630e18abbd9fa2a5b55f51ae7f60be585b898fc75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pxr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:49:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qc7hn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:17 crc kubenswrapper[4775]: I0321 04:50:17.952190 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbc8474b-4360-449d-ab37-ba14ca1ac5ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T04:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T04:48:33Z\\\",\\\"message\\\":\\\"le observer\\\\nW0321 04:48:33.168319 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 04:48:33.168481 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0321 04:48:33.169188 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3515854145/tls.crt::/tmp/serving-cert-3515854145/tls.key\\\\\\\"\\\\nI0321 04:48:33.461779 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 04:48:33.464865 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 04:48:33.464898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 04:48:33.464929 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 04:48:33.464945 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 04:48:33.471450 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 04:48:33.471478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471482 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 04:48:33.471486 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI0321 04:48:33.471486 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0321 04:48:33.471491 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 04:48:33.471520 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 04:48:33.471523 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0321 04:48:33.473616 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T04:48:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T04:47:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T04:47:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T04:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T04:47:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:17Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:18 crc kubenswrapper[4775]: I0321 04:50:18.660965 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:18 crc kubenswrapper[4775]: E0321 04:50:18.661096 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:18 crc kubenswrapper[4775]: E0321 04:50:18.767723 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.660745 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.660798 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.660825 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:19 crc kubenswrapper[4775]: E0321 04:50:19.661190 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:19 crc kubenswrapper[4775]: E0321 04:50:19.661289 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:19 crc kubenswrapper[4775]: E0321 04:50:19.661446 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.662256 4775 scope.go:117] "RemoveContainer" containerID="7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a" Mar 21 04:50:19 crc kubenswrapper[4775]: E0321 04:50:19.662399 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mzqtk_openshift-ovn-kubernetes(a69d31f5-deeb-4860-be96-ed5547831685)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" podUID="a69d31f5-deeb-4860-be96-ed5547831685" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.923893 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.923937 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.923949 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.923971 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.923983 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:50:19Z","lastTransitionTime":"2026-03-21T04:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:50:19 crc kubenswrapper[4775]: E0321 04:50:19.940459 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:19Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.946585 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.946677 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.946698 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.946776 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.946847 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:50:19Z","lastTransitionTime":"2026-03-21T04:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:50:19 crc kubenswrapper[4775]: E0321 04:50:19.962534 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:19Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:19 crc kubenswrapper[4775]: E0321 04:50:19.962965 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.962814 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs\") pod \"network-metrics-daemon-xk9f5\" (UID: \"6920413a-2c51-466d-a16e-d14489ae0c6c\") " pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:19 crc kubenswrapper[4775]: E0321 04:50:19.963037 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs podName:6920413a-2c51-466d-a16e-d14489ae0c6c nodeName:}" failed. No retries permitted until 2026-03-21 04:51:23.963016644 +0000 UTC m=+236.939480268 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs") pod "network-metrics-daemon-xk9f5" (UID: "6920413a-2c51-466d-a16e-d14489ae0c6c") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.968283 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.968319 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.968331 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.968349 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.968361 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:50:19Z","lastTransitionTime":"2026-03-21T04:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:50:19 crc kubenswrapper[4775]: E0321 04:50:19.984473 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:19Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.988836 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.988923 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.988941 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.988964 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:50:19 crc kubenswrapper[4775]: I0321 04:50:19.988983 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:50:19Z","lastTransitionTime":"2026-03-21T04:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:50:20 crc kubenswrapper[4775]: E0321 04:50:20.003452 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:20Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:20 crc kubenswrapper[4775]: I0321 04:50:20.007491 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:50:20 crc kubenswrapper[4775]: I0321 04:50:20.007548 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:50:20 crc kubenswrapper[4775]: I0321 04:50:20.007565 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:50:20 crc kubenswrapper[4775]: I0321 04:50:20.007586 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:50:20 crc kubenswrapper[4775]: I0321 04:50:20.007604 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:50:20Z","lastTransitionTime":"2026-03-21T04:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:50:20 crc kubenswrapper[4775]: E0321 04:50:20.027422 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T04:50:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"82de5c1b-41dd-43ea-8a7a-ca3ee73d6e07\\\",\\\"systemUUID\\\":\\\"b8f8f0e2-c78b-43d3-976e-2a86ca08a185\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T04:50:20Z is after 2025-08-24T17:21:41Z" Mar 21 04:50:20 crc kubenswrapper[4775]: E0321 04:50:20.027690 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 21 04:50:20 crc kubenswrapper[4775]: I0321 04:50:20.661364 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:20 crc kubenswrapper[4775]: E0321 04:50:20.661527 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:21 crc kubenswrapper[4775]: I0321 04:50:21.660929 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:21 crc kubenswrapper[4775]: E0321 04:50:21.661060 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:21 crc kubenswrapper[4775]: I0321 04:50:21.661240 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:21 crc kubenswrapper[4775]: I0321 04:50:21.661267 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:21 crc kubenswrapper[4775]: E0321 04:50:21.661738 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:21 crc kubenswrapper[4775]: E0321 04:50:21.661894 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:22 crc kubenswrapper[4775]: I0321 04:50:22.660716 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:22 crc kubenswrapper[4775]: E0321 04:50:22.661040 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:23 crc kubenswrapper[4775]: I0321 04:50:23.660586 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:23 crc kubenswrapper[4775]: I0321 04:50:23.660686 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:23 crc kubenswrapper[4775]: E0321 04:50:23.660723 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:23 crc kubenswrapper[4775]: E0321 04:50:23.660827 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:23 crc kubenswrapper[4775]: I0321 04:50:23.660851 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:23 crc kubenswrapper[4775]: E0321 04:50:23.661035 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:23 crc kubenswrapper[4775]: E0321 04:50:23.768531 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 04:50:24 crc kubenswrapper[4775]: I0321 04:50:24.661165 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:24 crc kubenswrapper[4775]: E0321 04:50:24.662227 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:25 crc kubenswrapper[4775]: I0321 04:50:25.660631 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:25 crc kubenswrapper[4775]: I0321 04:50:25.660641 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:25 crc kubenswrapper[4775]: E0321 04:50:25.660753 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:25 crc kubenswrapper[4775]: I0321 04:50:25.660708 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:25 crc kubenswrapper[4775]: E0321 04:50:25.660820 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:25 crc kubenswrapper[4775]: E0321 04:50:25.660922 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:26 crc kubenswrapper[4775]: I0321 04:50:26.661273 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:26 crc kubenswrapper[4775]: E0321 04:50:26.661425 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:27 crc kubenswrapper[4775]: I0321 04:50:27.660565 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:27 crc kubenswrapper[4775]: E0321 04:50:27.660730 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:27 crc kubenswrapper[4775]: I0321 04:50:27.660945 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:27 crc kubenswrapper[4775]: I0321 04:50:27.660982 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:27 crc kubenswrapper[4775]: E0321 04:50:27.661031 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:27 crc kubenswrapper[4775]: E0321 04:50:27.661190 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:27 crc kubenswrapper[4775]: I0321 04:50:27.688340 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=89.688325228 podStartE2EDuration="1m29.688325228s" podCreationTimestamp="2026-03-21 04:48:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:50:27.687913807 +0000 UTC m=+180.664377451" watchObservedRunningTime="2026-03-21 04:50:27.688325228 +0000 UTC m=+180.664788842" Mar 21 04:50:27 crc kubenswrapper[4775]: I0321 04:50:27.741450 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podStartSLOduration=121.741432741 podStartE2EDuration="2m1.741432741s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:50:27.726300804 +0000 UTC m=+180.702764428" watchObservedRunningTime="2026-03-21 04:50:27.741432741 +0000 UTC m=+180.717896365" Mar 21 04:50:27 crc kubenswrapper[4775]: I0321 04:50:27.772628 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=44.7726097 podStartE2EDuration="44.7726097s" podCreationTimestamp="2026-03-21 04:49:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:50:27.741600106 +0000 UTC m=+180.718063720" watchObservedRunningTime="2026-03-21 04:50:27.7726097 +0000 UTC m=+180.749073324" Mar 21 04:50:27 crc kubenswrapper[4775]: I0321 04:50:27.784942 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=94.784918416 podStartE2EDuration="1m34.784918416s" podCreationTimestamp="2026-03-21 04:48:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:50:27.772448196 +0000 UTC m=+180.748911820" watchObservedRunningTime="2026-03-21 04:50:27.784918416 +0000 UTC m=+180.761382040" Mar 21 04:50:27 crc kubenswrapper[4775]: I0321 04:50:27.818683 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=31.818667589 podStartE2EDuration="31.818667589s" podCreationTimestamp="2026-03-21 04:49:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:50:27.818405622 +0000 UTC m=+180.794869246" watchObservedRunningTime="2026-03-21 04:50:27.818667589 +0000 UTC m=+180.795131213" Mar 21 04:50:27 crc kubenswrapper[4775]: I0321 04:50:27.876872 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-556rg" podStartSLOduration=121.876841218 podStartE2EDuration="2m1.876841218s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:50:27.876798187 +0000 UTC m=+180.853261821" watchObservedRunningTime="2026-03-21 04:50:27.876841218 +0000 UTC m=+180.853304842" Mar 21 04:50:27 crc kubenswrapper[4775]: I0321 04:50:27.885112 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-kr988" podStartSLOduration=121.885093286 podStartE2EDuration="2m1.885093286s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:50:27.885074486 +0000 UTC m=+180.861538110" watchObservedRunningTime="2026-03-21 04:50:27.885093286 +0000 UTC m=+180.861556920" Mar 21 04:50:27 crc kubenswrapper[4775]: I0321 04:50:27.908730 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=33.908687757 podStartE2EDuration="33.908687757s" podCreationTimestamp="2026-03-21 04:49:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:50:27.907556734 +0000 UTC m=+180.884020358" watchObservedRunningTime="2026-03-21 04:50:27.908687757 +0000 UTC m=+180.885151401" Mar 21 04:50:27 crc kubenswrapper[4775]: I0321 04:50:27.919915 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-khh7x" podStartSLOduration=121.91989547 podStartE2EDuration="2m1.91989547s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:50:27.919154159 +0000 UTC m=+180.895617793" watchObservedRunningTime="2026-03-21 04:50:27.91989547 +0000 UTC m=+180.896359094" Mar 21 04:50:27 crc kubenswrapper[4775]: I0321 04:50:27.935456 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-kldzh" podStartSLOduration=121.935441159 podStartE2EDuration="2m1.935441159s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:50:27.934861632 +0000 UTC m=+180.911325256" watchObservedRunningTime="2026-03-21 04:50:27.935441159 +0000 UTC m=+180.911904783" Mar 21 04:50:27 crc kubenswrapper[4775]: I0321 04:50:27.950032 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-62jtz" podStartSLOduration=121.950014929 podStartE2EDuration="2m1.950014929s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:50:27.949067802 +0000 UTC m=+180.925531426" watchObservedRunningTime="2026-03-21 04:50:27.950014929 +0000 UTC m=+180.926478553" Mar 21 04:50:28 crc kubenswrapper[4775]: I0321 04:50:28.660371 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:28 crc kubenswrapper[4775]: E0321 04:50:28.660529 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:28 crc kubenswrapper[4775]: E0321 04:50:28.770306 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 04:50:29 crc kubenswrapper[4775]: I0321 04:50:29.661194 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:29 crc kubenswrapper[4775]: E0321 04:50:29.661344 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:29 crc kubenswrapper[4775]: I0321 04:50:29.661472 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:29 crc kubenswrapper[4775]: I0321 04:50:29.661592 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:29 crc kubenswrapper[4775]: E0321 04:50:29.661630 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:29 crc kubenswrapper[4775]: E0321 04:50:29.661826 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.378432 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.378482 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.378497 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.378514 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.378528 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T04:50:30Z","lastTransitionTime":"2026-03-21T04:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.424377 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8"] Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.424954 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.427361 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.427383 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.427431 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.427484 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.453694 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9758cbde-aaf6-4097-b10e-233d3efd75f5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fwjf8\" (UID: \"9758cbde-aaf6-4097-b10e-233d3efd75f5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.453744 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9758cbde-aaf6-4097-b10e-233d3efd75f5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fwjf8\" (UID: \"9758cbde-aaf6-4097-b10e-233d3efd75f5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.453799 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9758cbde-aaf6-4097-b10e-233d3efd75f5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fwjf8\" (UID: \"9758cbde-aaf6-4097-b10e-233d3efd75f5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.453875 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9758cbde-aaf6-4097-b10e-233d3efd75f5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fwjf8\" (UID: \"9758cbde-aaf6-4097-b10e-233d3efd75f5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.453932 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9758cbde-aaf6-4097-b10e-233d3efd75f5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fwjf8\" (UID: \"9758cbde-aaf6-4097-b10e-233d3efd75f5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.555055 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9758cbde-aaf6-4097-b10e-233d3efd75f5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fwjf8\" (UID: \"9758cbde-aaf6-4097-b10e-233d3efd75f5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.555190 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9758cbde-aaf6-4097-b10e-233d3efd75f5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fwjf8\" (UID: \"9758cbde-aaf6-4097-b10e-233d3efd75f5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.555259 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9758cbde-aaf6-4097-b10e-233d3efd75f5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fwjf8\" (UID: \"9758cbde-aaf6-4097-b10e-233d3efd75f5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.555352 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9758cbde-aaf6-4097-b10e-233d3efd75f5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fwjf8\" (UID: \"9758cbde-aaf6-4097-b10e-233d3efd75f5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.555405 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9758cbde-aaf6-4097-b10e-233d3efd75f5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fwjf8\" (UID: \"9758cbde-aaf6-4097-b10e-233d3efd75f5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.555496 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9758cbde-aaf6-4097-b10e-233d3efd75f5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fwjf8\" (UID: \"9758cbde-aaf6-4097-b10e-233d3efd75f5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.555552 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9758cbde-aaf6-4097-b10e-233d3efd75f5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fwjf8\" (UID: \"9758cbde-aaf6-4097-b10e-233d3efd75f5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.556373 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9758cbde-aaf6-4097-b10e-233d3efd75f5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fwjf8\" (UID: \"9758cbde-aaf6-4097-b10e-233d3efd75f5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.560894 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9758cbde-aaf6-4097-b10e-233d3efd75f5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fwjf8\" (UID: \"9758cbde-aaf6-4097-b10e-233d3efd75f5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.580857 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9758cbde-aaf6-4097-b10e-233d3efd75f5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fwjf8\" (UID: \"9758cbde-aaf6-4097-b10e-233d3efd75f5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.660615 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:30 crc kubenswrapper[4775]: E0321 04:50:30.660734 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.740388 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8" Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.786264 4775 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 21 04:50:30 crc kubenswrapper[4775]: I0321 04:50:30.794241 4775 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 21 04:50:31 crc kubenswrapper[4775]: I0321 04:50:31.091347 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8" event={"ID":"9758cbde-aaf6-4097-b10e-233d3efd75f5","Type":"ContainerStarted","Data":"f6c562abdcccdfd9d6df459e61ea1b3bee4017e7f28176b6a472e2488d95f2ae"} Mar 21 04:50:31 crc kubenswrapper[4775]: I0321 04:50:31.091400 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8" event={"ID":"9758cbde-aaf6-4097-b10e-233d3efd75f5","Type":"ContainerStarted","Data":"104c2205710b17c9c323460ac909eca961b1f4987df297698654484e316c551c"} Mar 21 04:50:31 crc kubenswrapper[4775]: I0321 04:50:31.103856 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fwjf8" podStartSLOduration=125.10383472 podStartE2EDuration="2m5.10383472s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:50:31.103713836 +0000 UTC m=+184.080177470" watchObservedRunningTime="2026-03-21 04:50:31.10383472 +0000 UTC m=+184.080298344" Mar 21 04:50:31 crc kubenswrapper[4775]: I0321 04:50:31.661201 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:31 crc kubenswrapper[4775]: I0321 04:50:31.661246 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:31 crc kubenswrapper[4775]: I0321 04:50:31.661201 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:31 crc kubenswrapper[4775]: E0321 04:50:31.661325 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:31 crc kubenswrapper[4775]: E0321 04:50:31.661396 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:31 crc kubenswrapper[4775]: E0321 04:50:31.661496 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:32 crc kubenswrapper[4775]: I0321 04:50:32.660879 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:32 crc kubenswrapper[4775]: E0321 04:50:32.661113 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:33 crc kubenswrapper[4775]: I0321 04:50:33.661397 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:33 crc kubenswrapper[4775]: I0321 04:50:33.661408 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:33 crc kubenswrapper[4775]: E0321 04:50:33.661554 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:33 crc kubenswrapper[4775]: E0321 04:50:33.661717 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:33 crc kubenswrapper[4775]: I0321 04:50:33.661840 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:33 crc kubenswrapper[4775]: E0321 04:50:33.662097 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:33 crc kubenswrapper[4775]: I0321 04:50:33.662488 4775 scope.go:117] "RemoveContainer" containerID="7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a" Mar 21 04:50:33 crc kubenswrapper[4775]: E0321 04:50:33.662692 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mzqtk_openshift-ovn-kubernetes(a69d31f5-deeb-4860-be96-ed5547831685)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" podUID="a69d31f5-deeb-4860-be96-ed5547831685" Mar 21 04:50:33 crc kubenswrapper[4775]: E0321 04:50:33.771504 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 04:50:34 crc kubenswrapper[4775]: I0321 04:50:34.660703 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:34 crc kubenswrapper[4775]: E0321 04:50:34.660838 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:35 crc kubenswrapper[4775]: I0321 04:50:35.661144 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:35 crc kubenswrapper[4775]: I0321 04:50:35.661173 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:35 crc kubenswrapper[4775]: I0321 04:50:35.661232 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:35 crc kubenswrapper[4775]: E0321 04:50:35.661527 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:35 crc kubenswrapper[4775]: E0321 04:50:35.661623 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:35 crc kubenswrapper[4775]: E0321 04:50:35.661361 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:36 crc kubenswrapper[4775]: I0321 04:50:36.661547 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:36 crc kubenswrapper[4775]: E0321 04:50:36.661707 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:37 crc kubenswrapper[4775]: I0321 04:50:37.661103 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:37 crc kubenswrapper[4775]: I0321 04:50:37.661209 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:37 crc kubenswrapper[4775]: I0321 04:50:37.661286 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:37 crc kubenswrapper[4775]: E0321 04:50:37.663065 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:37 crc kubenswrapper[4775]: E0321 04:50:37.663212 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:37 crc kubenswrapper[4775]: E0321 04:50:37.663405 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:38 crc kubenswrapper[4775]: I0321 04:50:38.660483 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:38 crc kubenswrapper[4775]: E0321 04:50:38.660989 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:38 crc kubenswrapper[4775]: E0321 04:50:38.772777 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 04:50:39 crc kubenswrapper[4775]: I0321 04:50:39.117033 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-556rg_e77ec218-42da-4f07-b214-184c4f3b20f3/kube-multus/1.log" Mar 21 04:50:39 crc kubenswrapper[4775]: I0321 04:50:39.117837 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-556rg_e77ec218-42da-4f07-b214-184c4f3b20f3/kube-multus/0.log" Mar 21 04:50:39 crc kubenswrapper[4775]: I0321 04:50:39.117904 4775 generic.go:334] "Generic (PLEG): container finished" podID="e77ec218-42da-4f07-b214-184c4f3b20f3" containerID="eaf83cd6f4a806265a871a5672f1a6939180308c2815b87ca254164f26d05aae" exitCode=1 Mar 21 04:50:39 crc kubenswrapper[4775]: I0321 04:50:39.117954 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-556rg" event={"ID":"e77ec218-42da-4f07-b214-184c4f3b20f3","Type":"ContainerDied","Data":"eaf83cd6f4a806265a871a5672f1a6939180308c2815b87ca254164f26d05aae"} Mar 21 04:50:39 crc kubenswrapper[4775]: I0321 04:50:39.118008 4775 scope.go:117] "RemoveContainer" containerID="77509499e5f9ecba799d75a9e9f99a1b2ad760f16f11a518f1d8bd7f848acabf" Mar 21 04:50:39 crc kubenswrapper[4775]: I0321 04:50:39.118526 4775 scope.go:117] "RemoveContainer" containerID="eaf83cd6f4a806265a871a5672f1a6939180308c2815b87ca254164f26d05aae" Mar 21 04:50:39 crc kubenswrapper[4775]: E0321 04:50:39.118709 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-556rg_openshift-multus(e77ec218-42da-4f07-b214-184c4f3b20f3)\"" pod="openshift-multus/multus-556rg" podUID="e77ec218-42da-4f07-b214-184c4f3b20f3" Mar 21 04:50:39 crc kubenswrapper[4775]: I0321 04:50:39.661500 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:39 crc kubenswrapper[4775]: I0321 04:50:39.661611 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:39 crc kubenswrapper[4775]: I0321 04:50:39.661964 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:39 crc kubenswrapper[4775]: E0321 04:50:39.662091 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:39 crc kubenswrapper[4775]: E0321 04:50:39.662293 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:39 crc kubenswrapper[4775]: E0321 04:50:39.662383 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:40 crc kubenswrapper[4775]: I0321 04:50:40.121998 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-556rg_e77ec218-42da-4f07-b214-184c4f3b20f3/kube-multus/1.log" Mar 21 04:50:40 crc kubenswrapper[4775]: I0321 04:50:40.660898 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:40 crc kubenswrapper[4775]: E0321 04:50:40.661065 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:41 crc kubenswrapper[4775]: I0321 04:50:41.660505 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:41 crc kubenswrapper[4775]: I0321 04:50:41.660592 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:41 crc kubenswrapper[4775]: E0321 04:50:41.660664 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:41 crc kubenswrapper[4775]: I0321 04:50:41.660499 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:41 crc kubenswrapper[4775]: E0321 04:50:41.660912 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:41 crc kubenswrapper[4775]: E0321 04:50:41.660976 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:42 crc kubenswrapper[4775]: I0321 04:50:42.660331 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:42 crc kubenswrapper[4775]: E0321 04:50:42.660535 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:43 crc kubenswrapper[4775]: I0321 04:50:43.660800 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:43 crc kubenswrapper[4775]: I0321 04:50:43.660840 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:43 crc kubenswrapper[4775]: E0321 04:50:43.661065 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:43 crc kubenswrapper[4775]: E0321 04:50:43.661097 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:43 crc kubenswrapper[4775]: I0321 04:50:43.660861 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:43 crc kubenswrapper[4775]: E0321 04:50:43.661196 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:43 crc kubenswrapper[4775]: E0321 04:50:43.774668 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 04:50:44 crc kubenswrapper[4775]: I0321 04:50:44.661315 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:44 crc kubenswrapper[4775]: E0321 04:50:44.661442 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:45 crc kubenswrapper[4775]: I0321 04:50:45.660920 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:45 crc kubenswrapper[4775]: I0321 04:50:45.661193 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:45 crc kubenswrapper[4775]: I0321 04:50:45.661210 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:45 crc kubenswrapper[4775]: E0321 04:50:45.661573 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:45 crc kubenswrapper[4775]: E0321 04:50:45.661655 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:45 crc kubenswrapper[4775]: E0321 04:50:45.661687 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:46 crc kubenswrapper[4775]: I0321 04:50:46.660321 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:46 crc kubenswrapper[4775]: E0321 04:50:46.660707 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:47 crc kubenswrapper[4775]: I0321 04:50:47.660299 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:47 crc kubenswrapper[4775]: E0321 04:50:47.661633 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:47 crc kubenswrapper[4775]: I0321 04:50:47.661677 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:47 crc kubenswrapper[4775]: I0321 04:50:47.661733 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:47 crc kubenswrapper[4775]: E0321 04:50:47.661877 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:47 crc kubenswrapper[4775]: E0321 04:50:47.661939 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:48 crc kubenswrapper[4775]: I0321 04:50:48.661501 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:48 crc kubenswrapper[4775]: E0321 04:50:48.662024 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:48 crc kubenswrapper[4775]: I0321 04:50:48.662279 4775 scope.go:117] "RemoveContainer" containerID="7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a" Mar 21 04:50:48 crc kubenswrapper[4775]: E0321 04:50:48.775943 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 04:50:49 crc kubenswrapper[4775]: I0321 04:50:49.182066 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzqtk_a69d31f5-deeb-4860-be96-ed5547831685/ovnkube-controller/3.log" Mar 21 04:50:49 crc kubenswrapper[4775]: I0321 04:50:49.184581 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerStarted","Data":"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646"} Mar 21 04:50:49 crc kubenswrapper[4775]: I0321 04:50:49.185033 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:50:49 crc kubenswrapper[4775]: I0321 04:50:49.213866 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" podStartSLOduration=143.213839381 podStartE2EDuration="2m23.213839381s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:50:49.211452582 +0000 UTC m=+202.187916206" watchObservedRunningTime="2026-03-21 04:50:49.213839381 +0000 UTC m=+202.190303005" Mar 21 04:50:49 crc kubenswrapper[4775]: I0321 04:50:49.519295 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-xk9f5"] Mar 21 04:50:49 crc kubenswrapper[4775]: I0321 04:50:49.519411 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:49 crc kubenswrapper[4775]: E0321 04:50:49.519486 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:49 crc kubenswrapper[4775]: I0321 04:50:49.661218 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:49 crc kubenswrapper[4775]: I0321 04:50:49.661293 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:49 crc kubenswrapper[4775]: E0321 04:50:49.661336 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:49 crc kubenswrapper[4775]: E0321 04:50:49.661485 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:50 crc kubenswrapper[4775]: I0321 04:50:50.661100 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:50 crc kubenswrapper[4775]: E0321 04:50:50.661267 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:51 crc kubenswrapper[4775]: I0321 04:50:51.661312 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:51 crc kubenswrapper[4775]: I0321 04:50:51.661318 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:51 crc kubenswrapper[4775]: E0321 04:50:51.661573 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:51 crc kubenswrapper[4775]: I0321 04:50:51.661611 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:51 crc kubenswrapper[4775]: E0321 04:50:51.661829 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:51 crc kubenswrapper[4775]: E0321 04:50:51.661906 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:52 crc kubenswrapper[4775]: I0321 04:50:52.661014 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:52 crc kubenswrapper[4775]: E0321 04:50:52.661353 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:53 crc kubenswrapper[4775]: I0321 04:50:53.660681 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:53 crc kubenswrapper[4775]: I0321 04:50:53.660786 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:53 crc kubenswrapper[4775]: I0321 04:50:53.660825 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:53 crc kubenswrapper[4775]: E0321 04:50:53.660966 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:53 crc kubenswrapper[4775]: E0321 04:50:53.661211 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:53 crc kubenswrapper[4775]: I0321 04:50:53.661221 4775 scope.go:117] "RemoveContainer" containerID="eaf83cd6f4a806265a871a5672f1a6939180308c2815b87ca254164f26d05aae" Mar 21 04:50:53 crc kubenswrapper[4775]: E0321 04:50:53.661266 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:53 crc kubenswrapper[4775]: E0321 04:50:53.778144 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 04:50:54 crc kubenswrapper[4775]: I0321 04:50:54.213641 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-556rg_e77ec218-42da-4f07-b214-184c4f3b20f3/kube-multus/1.log" Mar 21 04:50:54 crc kubenswrapper[4775]: I0321 04:50:54.213713 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-556rg" event={"ID":"e77ec218-42da-4f07-b214-184c4f3b20f3","Type":"ContainerStarted","Data":"ad12949f26afe1756d4a6c0d01069cb26a8928bb36cc24602d4ab1bbde117f9e"} Mar 21 04:50:54 crc kubenswrapper[4775]: I0321 04:50:54.660534 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:54 crc kubenswrapper[4775]: E0321 04:50:54.660753 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:55 crc kubenswrapper[4775]: I0321 04:50:55.660927 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:55 crc kubenswrapper[4775]: I0321 04:50:55.660927 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:55 crc kubenswrapper[4775]: I0321 04:50:55.660987 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:55 crc kubenswrapper[4775]: E0321 04:50:55.662009 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:55 crc kubenswrapper[4775]: E0321 04:50:55.662369 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:55 crc kubenswrapper[4775]: E0321 04:50:55.662405 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:56 crc kubenswrapper[4775]: I0321 04:50:56.660758 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:56 crc kubenswrapper[4775]: E0321 04:50:56.660897 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:57 crc kubenswrapper[4775]: I0321 04:50:57.660936 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:57 crc kubenswrapper[4775]: I0321 04:50:57.660946 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:57 crc kubenswrapper[4775]: E0321 04:50:57.663392 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 04:50:57 crc kubenswrapper[4775]: I0321 04:50:57.663646 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:57 crc kubenswrapper[4775]: E0321 04:50:57.663742 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xk9f5" podUID="6920413a-2c51-466d-a16e-d14489ae0c6c" Mar 21 04:50:57 crc kubenswrapper[4775]: E0321 04:50:57.663926 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 04:50:58 crc kubenswrapper[4775]: I0321 04:50:58.660968 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:50:58 crc kubenswrapper[4775]: E0321 04:50:58.661102 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 04:50:59 crc kubenswrapper[4775]: I0321 04:50:59.661388 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:50:59 crc kubenswrapper[4775]: I0321 04:50:59.661490 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:50:59 crc kubenswrapper[4775]: I0321 04:50:59.661439 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:50:59 crc kubenswrapper[4775]: I0321 04:50:59.664380 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 21 04:50:59 crc kubenswrapper[4775]: I0321 04:50:59.664896 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 21 04:50:59 crc kubenswrapper[4775]: I0321 04:50:59.665302 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 21 04:50:59 crc kubenswrapper[4775]: I0321 04:50:59.667764 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.418338 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:00 crc kubenswrapper[4775]: E0321 04:51:00.418616 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:53:02.418581556 +0000 UTC m=+335.395045180 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.419933 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.419989 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:51:00 crc kubenswrapper[4775]: E0321 04:51:00.420176 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 04:51:00 crc kubenswrapper[4775]: E0321 04:51:00.420224 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 04:53:02.420211963 +0000 UTC m=+335.396675587 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 04:51:00 crc kubenswrapper[4775]: E0321 04:51:00.420513 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 04:51:00 crc kubenswrapper[4775]: E0321 04:51:00.420678 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 04:53:02.420656776 +0000 UTC m=+335.397120400 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.520732 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.520779 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.528057 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.529800 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.582172 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.600988 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.660353 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.663526 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.665008 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.831869 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.877227 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k5dks"] Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.877890 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k5dks" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.878724 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dzxd5"] Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.878961 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf"] Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.879277 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mthhs"] Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.879308 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.879549 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.879597 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-dzxd5" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.881048 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-fc4z8"] Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.881467 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jd86b"] Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.881920 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.882354 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-fc4z8" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.885854 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.886548 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r6c6z"] Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.886937 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.887345 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.887500 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rz6g5"] Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.888053 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.888219 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-wst2s"] Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.888638 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.889214 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.889460 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.899028 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.900754 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.900931 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5rnj6"] Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.901436 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5pdcr"] Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.901755 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5pdcr" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.902014 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-5rnj6" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.902362 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.902583 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.902700 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.902724 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.902803 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.902818 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.902823 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.902888 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.903038 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.903136 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.903168 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.903245 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.903356 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.903379 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.903408 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.903505 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.903591 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.903739 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.903940 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.904225 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.904272 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.904304 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.904378 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.904412 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.904441 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.904505 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.904532 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.904593 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.904645 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.904700 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.904725 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.904792 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.905000 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.905022 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.905053 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.905096 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.904507 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.905204 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.904505 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.905275 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.905178 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.905345 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.905423 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.905441 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.905573 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.905656 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.905723 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.905776 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp"] Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.906146 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.906426 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-nstnr"] Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.907136 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.907870 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9v866"] Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.908102 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-nstnr" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.908193 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9v866" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.909257 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.909267 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.911511 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.912202 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.912611 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.912710 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.912837 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.913040 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.913805 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-468g4"] Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.914421 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-468g4" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.921348 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.923716 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.923864 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.923924 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.923999 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.924018 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.924477 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.924591 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.924799 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.924848 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.924953 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.927193 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 21 04:51:00 crc kubenswrapper[4775]: I0321 04:51:00.927275 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.052086 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.052558 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.053227 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-57cxh"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.053300 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.053336 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6e22b5c7-4191-4f21-82ba-3014ccc4e978-etcd-client\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.053364 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6e22b5c7-4191-4f21-82ba-3014ccc4e978-encryption-config\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.053770 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-57cxh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.054078 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.055013 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7sfkh"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.055536 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zkl7p"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.055857 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zkl7p" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.055878 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7sfkh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056039 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12b1cc27-6a60-43d2-9d5e-eb7a54c1e899-config\") pod \"console-operator-58897d9998-fc4z8\" (UID: \"12b1cc27-6a60-43d2-9d5e-eb7a54c1e899\") " pod="openshift-console-operator/console-operator-58897d9998-fc4z8" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.055859 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056093 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/fe3df1e1-4c22-48df-aaea-469c864f0310-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5rnj6\" (UID: \"fe3df1e1-4c22-48df-aaea-469c864f0310\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5rnj6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056158 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e22b5c7-4191-4f21-82ba-3014ccc4e978-config\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056187 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nbfq\" (UniqueName: \"kubernetes.io/projected/2d2512b8-9fb6-4b9b-9ac2-268e0b8bd04a-kube-api-access-7nbfq\") pod \"openshift-apiserver-operator-796bbdcf4f-5pdcr\" (UID: \"2d2512b8-9fb6-4b9b-9ac2-268e0b8bd04a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5pdcr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056218 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-trusted-ca-bundle\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056247 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-oauth-serving-cert\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056271 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqkw4\" (UniqueName: \"kubernetes.io/projected/8bb7828f-6d99-4539-8312-c8e96bfbc608-kube-api-access-lqkw4\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056298 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2lnkp"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056368 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-client-ca\") pod \"route-controller-manager-6576b87f9c-55lcp\" (UID: \"ed9baf4c-70b2-450f-9b21-76dfafbc44d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056438 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-config\") pod \"controller-manager-879f6c89f-r6c6z\" (UID: \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056519 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6e22b5c7-4191-4f21-82ba-3014ccc4e978-audit-dir\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056562 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/edb4245a-7971-4c40-81b6-27d56b319a2f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-k5dks\" (UID: \"edb4245a-7971-4c40-81b6-27d56b319a2f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k5dks" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056623 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056651 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/12b1cc27-6a60-43d2-9d5e-eb7a54c1e899-trusted-ca\") pod \"console-operator-58897d9998-fc4z8\" (UID: \"12b1cc27-6a60-43d2-9d5e-eb7a54c1e899\") " pod="openshift-console-operator/console-operator-58897d9998-fc4z8" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056689 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/892d15b6-460e-4892-a836-0cc284c8a326-serving-cert\") pod \"openshift-config-operator-7777fb866f-mthhs\" (UID: \"892d15b6-460e-4892-a836-0cc284c8a326\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056717 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056796 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056842 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-service-ca\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056865 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/10a9b5d9-308c-4971-a073-9c88de98d8ea-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9v866\" (UID: \"10a9b5d9-308c-4971-a073-9c88de98d8ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9v866" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056887 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12b1cc27-6a60-43d2-9d5e-eb7a54c1e899-serving-cert\") pod \"console-operator-58897d9998-fc4z8\" (UID: \"12b1cc27-6a60-43d2-9d5e-eb7a54c1e899\") " pod="openshift-console-operator/console-operator-58897d9998-fc4z8" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056921 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/768c9343-1391-4221-b008-5dee2921953f-config\") pod \"authentication-operator-69f744f599-dzxd5\" (UID: \"768c9343-1391-4221-b008-5dee2921953f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dzxd5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056941 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8bb7828f-6d99-4539-8312-c8e96bfbc608-audit-dir\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056962 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k97vh\" (UniqueName: \"kubernetes.io/projected/fe3df1e1-4c22-48df-aaea-469c864f0310-kube-api-access-k97vh\") pod \"machine-api-operator-5694c8668f-5rnj6\" (UID: \"fe3df1e1-4c22-48df-aaea-469c864f0310\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5rnj6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056977 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/768c9343-1391-4221-b008-5dee2921953f-service-ca-bundle\") pod \"authentication-operator-69f744f599-dzxd5\" (UID: \"768c9343-1391-4221-b008-5dee2921953f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dzxd5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.056992 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6e22b5c7-4191-4f21-82ba-3014ccc4e978-node-pullsecrets\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057010 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fddde3da-8512-4e62-9c38-b59f98e117e0-audit-dir\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057027 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhsgq\" (UniqueName: \"kubernetes.io/projected/c3d58eba-4ddf-463c-baa1-1943fb60c732-kube-api-access-bhsgq\") pod \"downloads-7954f5f757-nstnr\" (UID: \"c3d58eba-4ddf-463c-baa1-1943fb60c732\") " pod="openshift-console/downloads-7954f5f757-nstnr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057041 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntrqw\" (UniqueName: \"kubernetes.io/projected/768c9343-1391-4221-b008-5dee2921953f-kube-api-access-ntrqw\") pod \"authentication-operator-69f744f599-dzxd5\" (UID: \"768c9343-1391-4221-b008-5dee2921953f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dzxd5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057058 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zb5h\" (UniqueName: \"kubernetes.io/projected/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-kube-api-access-2zb5h\") pod \"controller-manager-879f6c89f-r6c6z\" (UID: \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057095 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-2lnkp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057131 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-console-serving-cert\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057171 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/10a9b5d9-308c-4971-a073-9c88de98d8ea-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9v866\" (UID: \"10a9b5d9-308c-4971-a073-9c88de98d8ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9v866" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057199 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6343dd4f-5c5b-4c94-a4d6-f603698ba6ac-machine-approver-tls\") pod \"machine-approver-56656f9798-468g4\" (UID: \"6343dd4f-5c5b-4c94-a4d6-f603698ba6ac\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-468g4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057221 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8bb7828f-6d99-4539-8312-c8e96bfbc608-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057247 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6e22b5c7-4191-4f21-82ba-3014ccc4e978-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057275 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhtwp\" (UniqueName: \"kubernetes.io/projected/892d15b6-460e-4892-a836-0cc284c8a326-kube-api-access-mhtwp\") pod \"openshift-config-operator-7777fb866f-mthhs\" (UID: \"892d15b6-460e-4892-a836-0cc284c8a326\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057297 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pflfp\" (UniqueName: \"kubernetes.io/projected/6343dd4f-5c5b-4c94-a4d6-f603698ba6ac-kube-api-access-pflfp\") pod \"machine-approver-56656f9798-468g4\" (UID: \"6343dd4f-5c5b-4c94-a4d6-f603698ba6ac\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-468g4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057322 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8bb7828f-6d99-4539-8312-c8e96bfbc608-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057343 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-client-ca\") pod \"controller-manager-879f6c89f-r6c6z\" (UID: \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057369 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-serving-cert\") pod \"route-controller-manager-6576b87f9c-55lcp\" (UID: \"ed9baf4c-70b2-450f-9b21-76dfafbc44d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057390 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e22b5c7-4191-4f21-82ba-3014ccc4e978-serving-cert\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057401 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-725jh"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057414 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057439 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057462 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6343dd4f-5c5b-4c94-a4d6-f603698ba6ac-config\") pod \"machine-approver-56656f9798-468g4\" (UID: \"6343dd4f-5c5b-4c94-a4d6-f603698ba6ac\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-468g4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057482 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8bb7828f-6d99-4539-8312-c8e96bfbc608-serving-cert\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057504 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvbpw\" (UniqueName: \"kubernetes.io/projected/edb4245a-7971-4c40-81b6-27d56b319a2f-kube-api-access-dvbpw\") pod \"cluster-samples-operator-665b6dd947-k5dks\" (UID: \"edb4245a-7971-4c40-81b6-27d56b319a2f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k5dks" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057526 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057549 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/768c9343-1391-4221-b008-5dee2921953f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dzxd5\" (UID: \"768c9343-1391-4221-b008-5dee2921953f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dzxd5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057576 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fw6r\" (UniqueName: \"kubernetes.io/projected/fddde3da-8512-4e62-9c38-b59f98e117e0-kube-api-access-4fw6r\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057598 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m6fm\" (UniqueName: \"kubernetes.io/projected/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-kube-api-access-8m6fm\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057619 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/892d15b6-460e-4892-a836-0cc284c8a326-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mthhs\" (UID: \"892d15b6-460e-4892-a836-0cc284c8a326\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057678 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8bb7828f-6d99-4539-8312-c8e96bfbc608-audit-policies\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057703 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/10a9b5d9-308c-4971-a073-9c88de98d8ea-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9v866\" (UID: \"10a9b5d9-308c-4971-a073-9c88de98d8ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9v866" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057724 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8bb7828f-6d99-4539-8312-c8e96bfbc608-etcd-client\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057746 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d2512b8-9fb6-4b9b-9ac2-268e0b8bd04a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5pdcr\" (UID: \"2d2512b8-9fb6-4b9b-9ac2-268e0b8bd04a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5pdcr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057777 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7h98\" (UniqueName: \"kubernetes.io/projected/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-kube-api-access-r7h98\") pod \"route-controller-manager-6576b87f9c-55lcp\" (UID: \"ed9baf4c-70b2-450f-9b21-76dfafbc44d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057797 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d2512b8-9fb6-4b9b-9ac2-268e0b8bd04a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5pdcr\" (UID: \"2d2512b8-9fb6-4b9b-9ac2-268e0b8bd04a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5pdcr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057819 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk47r\" (UniqueName: \"kubernetes.io/projected/12b1cc27-6a60-43d2-9d5e-eb7a54c1e899-kube-api-access-rk47r\") pod \"console-operator-58897d9998-fc4z8\" (UID: \"12b1cc27-6a60-43d2-9d5e-eb7a54c1e899\") " pod="openshift-console-operator/console-operator-58897d9998-fc4z8" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057846 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-console-config\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057866 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8bb7828f-6d99-4539-8312-c8e96bfbc608-encryption-config\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057875 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-725jh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057888 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-console-oauth-config\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057910 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6e22b5c7-4191-4f21-82ba-3014ccc4e978-image-import-ca\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057932 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe3df1e1-4c22-48df-aaea-469c864f0310-config\") pod \"machine-api-operator-5694c8668f-5rnj6\" (UID: \"fe3df1e1-4c22-48df-aaea-469c864f0310\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5rnj6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057954 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fe3df1e1-4c22-48df-aaea-469c864f0310-images\") pod \"machine-api-operator-5694c8668f-5rnj6\" (UID: \"fe3df1e1-4c22-48df-aaea-469c864f0310\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5rnj6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.057979 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td4ls\" (UniqueName: \"kubernetes.io/projected/10a9b5d9-308c-4971-a073-9c88de98d8ea-kube-api-access-td4ls\") pod \"cluster-image-registry-operator-dc59b4c8b-9v866\" (UID: \"10a9b5d9-308c-4971-a073-9c88de98d8ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9v866" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.058000 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6343dd4f-5c5b-4c94-a4d6-f603698ba6ac-auth-proxy-config\") pod \"machine-approver-56656f9798-468g4\" (UID: \"6343dd4f-5c5b-4c94-a4d6-f603698ba6ac\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-468g4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.058021 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-config\") pod \"route-controller-manager-6576b87f9c-55lcp\" (UID: \"ed9baf4c-70b2-450f-9b21-76dfafbc44d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.058043 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6e22b5c7-4191-4f21-82ba-3014ccc4e978-audit\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.058067 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.058088 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.058111 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/768c9343-1391-4221-b008-5dee2921953f-serving-cert\") pod \"authentication-operator-69f744f599-dzxd5\" (UID: \"768c9343-1391-4221-b008-5dee2921953f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dzxd5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.058155 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-serving-cert\") pod \"controller-manager-879f6c89f-r6c6z\" (UID: \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.058178 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-audit-policies\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.058227 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.058366 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.058388 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6e22b5c7-4191-4f21-82ba-3014ccc4e978-etcd-serving-ca\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.058412 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-r6c6z\" (UID: \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.058432 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75ck2\" (UniqueName: \"kubernetes.io/projected/6e22b5c7-4191-4f21-82ba-3014ccc4e978-kube-api-access-75ck2\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.058564 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hvd8t"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.059450 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.060511 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.061402 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-d7xpn"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.061480 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.061790 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-d7xpn" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.061957 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.061975 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.064018 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.064208 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.064304 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.064458 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.064752 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.065128 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-blwmr"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.065778 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-blwmr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.066939 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mbgvc"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.067636 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mbgvc" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.067664 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.068154 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.068256 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.068505 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.070472 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.072774 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.073078 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.073329 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.074201 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v296h"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.074798 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v296h" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.075068 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-gd4gc"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.075745 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-gd4gc" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.077373 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.077413 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.078328 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.078671 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.079130 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.093455 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.095455 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.093946 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wrzs4"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.098151 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-4nfs6"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.099073 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.100437 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.102486 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.105590 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.106591 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.107590 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.108736 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.108982 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vjv4q"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.109265 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.109940 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vjv4q" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.111260 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wqjnj"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.111924 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567810-wb89g"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.112185 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wqjnj" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.112596 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567810-wb89g" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.115019 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.115708 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.115792 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.116401 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.116794 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.117050 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.117610 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-2z46g"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.118823 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.119575 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zh8p"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.120230 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zh8p" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.120579 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.122803 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xv7nw"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.123669 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xv7nw" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.127815 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5ztsd"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.128499 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5ztsd" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.129468 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.130840 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.131185 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-zs5pr"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.131991 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-zs5pr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.132603 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mthhs"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.134007 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k5dks"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.135350 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dzxd5"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.136837 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-hvfsp"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.138280 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.139194 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.140383 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.140777 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-gd4gc"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.142495 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r6c6z"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.143779 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9v866"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.144981 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jd86b"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.146227 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-fc4z8"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.147091 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-d7xpn"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.148146 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-4nfs6"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.149139 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-57cxh"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.150193 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-wst2s"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.151197 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7sfkh"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.152268 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5rnj6"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.153315 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hvd8t"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.154533 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-nstnr"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.155658 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mbgvc"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.156824 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5pdcr"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.158123 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v296h"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.158846 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dfbaac71-f99c-4373-a469-f2e5dd0ee632-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wrzs4\" (UID: \"dfbaac71-f99c-4373-a469-f2e5dd0ee632\") " pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.158942 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f9395e24-0d4b-4165-bf60-068876927f58-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5ztsd\" (UID: \"f9395e24-0d4b-4165-bf60-068876927f58\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5ztsd" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.159022 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/12b1cc27-6a60-43d2-9d5e-eb7a54c1e899-trusted-ca\") pod \"console-operator-58897d9998-fc4z8\" (UID: \"12b1cc27-6a60-43d2-9d5e-eb7a54c1e899\") " pod="openshift-console-operator/console-operator-58897d9998-fc4z8" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.159099 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07720715-40fb-4f7d-8f6b-381dc7cf9ea2-serving-cert\") pod \"etcd-operator-b45778765-4nfs6\" (UID: \"07720715-40fb-4f7d-8f6b-381dc7cf9ea2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.159213 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.159387 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae87529c-e52f-45b0-9bbf-2a652e628bc5-metrics-tls\") pod \"dns-default-zs5pr\" (UID: \"ae87529c-e52f-45b0-9bbf-2a652e628bc5\") " pod="openshift-dns/dns-default-zs5pr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.159477 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clmnm\" (UniqueName: \"kubernetes.io/projected/9de83f0b-7dd2-4846-a1ce-c8af930778f4-kube-api-access-clmnm\") pod \"csi-hostpathplugin-hvfsp\" (UID: \"9de83f0b-7dd2-4846-a1ce-c8af930778f4\") " pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.159620 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12b1cc27-6a60-43d2-9d5e-eb7a54c1e899-serving-cert\") pod \"console-operator-58897d9998-fc4z8\" (UID: \"12b1cc27-6a60-43d2-9d5e-eb7a54c1e899\") " pod="openshift-console-operator/console-operator-58897d9998-fc4z8" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.159701 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eddac2a3-8b07-410d-9e9f-f7ce0c05abf5-config\") pod \"kube-controller-manager-operator-78b949d7b-725jh\" (UID: \"eddac2a3-8b07-410d-9e9f-f7ce0c05abf5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-725jh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.159570 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vjv4q"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.159869 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c0dc3b9-8e11-4ff0-8f26-549ce47215f0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-blwmr\" (UID: \"0c0dc3b9-8e11-4ff0-8f26-549ce47215f0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-blwmr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.159988 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/768c9343-1391-4221-b008-5dee2921953f-config\") pod \"authentication-operator-69f744f599-dzxd5\" (UID: \"768c9343-1391-4221-b008-5dee2921953f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dzxd5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160019 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9mb9\" (UniqueName: \"kubernetes.io/projected/edb67e99-0d50-46e9-adbb-c0831dd915d8-kube-api-access-d9mb9\") pod \"openshift-controller-manager-operator-756b6f6bc6-xv7nw\" (UID: \"edb67e99-0d50-46e9-adbb-c0831dd915d8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xv7nw" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160037 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a79bbac6-f40a-4c92-8854-7ab5e72573cc-stats-auth\") pod \"router-default-5444994796-2z46g\" (UID: \"a79bbac6-f40a-4c92-8854-7ab5e72573cc\") " pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160056 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nw58z\" (UniqueName: \"kubernetes.io/projected/07720715-40fb-4f7d-8f6b-381dc7cf9ea2-kube-api-access-nw58z\") pod \"etcd-operator-b45778765-4nfs6\" (UID: \"07720715-40fb-4f7d-8f6b-381dc7cf9ea2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.159769 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160075 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k97vh\" (UniqueName: \"kubernetes.io/projected/fe3df1e1-4c22-48df-aaea-469c864f0310-kube-api-access-k97vh\") pod \"machine-api-operator-5694c8668f-5rnj6\" (UID: \"fe3df1e1-4c22-48df-aaea-469c864f0310\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5rnj6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160183 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6e22b5c7-4191-4f21-82ba-3014ccc4e978-node-pullsecrets\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160211 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhsgq\" (UniqueName: \"kubernetes.io/projected/c3d58eba-4ddf-463c-baa1-1943fb60c732-kube-api-access-bhsgq\") pod \"downloads-7954f5f757-nstnr\" (UID: \"c3d58eba-4ddf-463c-baa1-1943fb60c732\") " pod="openshift-console/downloads-7954f5f757-nstnr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160229 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6e22b5c7-4191-4f21-82ba-3014ccc4e978-node-pullsecrets\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160234 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntrqw\" (UniqueName: \"kubernetes.io/projected/768c9343-1391-4221-b008-5dee2921953f-kube-api-access-ntrqw\") pod \"authentication-operator-69f744f599-dzxd5\" (UID: \"768c9343-1391-4221-b008-5dee2921953f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dzxd5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160256 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zb5h\" (UniqueName: \"kubernetes.io/projected/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-kube-api-access-2zb5h\") pod \"controller-manager-879f6c89f-r6c6z\" (UID: \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160298 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a49709c7-59e0-440e-89c2-177c42cd28e8-tmpfs\") pod \"packageserver-d55dfcdfc-sp69c\" (UID: \"a49709c7-59e0-440e-89c2-177c42cd28e8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160318 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zfpw\" (UniqueName: \"kubernetes.io/projected/e5224539-6d29-4bc3-9656-4665eb287e28-kube-api-access-7zfpw\") pod \"control-plane-machine-set-operator-78cbb6b69f-d7xpn\" (UID: \"e5224539-6d29-4bc3-9656-4665eb287e28\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-d7xpn" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160361 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plx8q\" (UniqueName: \"kubernetes.io/projected/b0b465ff-f8f3-4a99-9235-1ddc4ce093e4-kube-api-access-plx8q\") pod \"olm-operator-6b444d44fb-v296h\" (UID: \"b0b465ff-f8f3-4a99-9235-1ddc4ce093e4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v296h" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160391 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-console-serving-cert\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160441 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160456 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/68d61e65-8275-4862-9dae-a75029889b2a-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-57cxh\" (UID: \"68d61e65-8275-4862-9dae-a75029889b2a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-57cxh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160519 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/811f064c-ebf4-48ad-87a0-83205eb1eca5-metrics-tls\") pod \"dns-operator-744455d44c-vjv4q\" (UID: \"811f064c-ebf4-48ad-87a0-83205eb1eca5\") " pod="openshift-dns-operator/dns-operator-744455d44c-vjv4q" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160547 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a49709c7-59e0-440e-89c2-177c42cd28e8-webhook-cert\") pod \"packageserver-d55dfcdfc-sp69c\" (UID: \"a49709c7-59e0-440e-89c2-177c42cd28e8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160570 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6e22b5c7-4191-4f21-82ba-3014ccc4e978-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160684 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pflfp\" (UniqueName: \"kubernetes.io/projected/6343dd4f-5c5b-4c94-a4d6-f603698ba6ac-kube-api-access-pflfp\") pod \"machine-approver-56656f9798-468g4\" (UID: \"6343dd4f-5c5b-4c94-a4d6-f603698ba6ac\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-468g4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160713 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-serving-cert\") pod \"route-controller-manager-6576b87f9c-55lcp\" (UID: \"ed9baf4c-70b2-450f-9b21-76dfafbc44d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160736 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhtwp\" (UniqueName: \"kubernetes.io/projected/892d15b6-460e-4892-a836-0cc284c8a326-kube-api-access-mhtwp\") pod \"openshift-config-operator-7777fb866f-mthhs\" (UID: \"892d15b6-460e-4892-a836-0cc284c8a326\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160786 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zkl7p"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160796 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e22b5c7-4191-4f21-82ba-3014ccc4e978-serving-cert\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160842 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160869 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gr9p\" (UniqueName: \"kubernetes.io/projected/2dbc150e-3a25-4b44-b01c-effe99de5152-kube-api-access-5gr9p\") pod \"migrator-59844c95c7-7sfkh\" (UID: \"2dbc150e-3a25-4b44-b01c-effe99de5152\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7sfkh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160889 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a79bbac6-f40a-4c92-8854-7ab5e72573cc-default-certificate\") pod \"router-default-5444994796-2z46g\" (UID: \"a79bbac6-f40a-4c92-8854-7ab5e72573cc\") " pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160940 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/768c9343-1391-4221-b008-5dee2921953f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dzxd5\" (UID: \"768c9343-1391-4221-b008-5dee2921953f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dzxd5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160967 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fw6r\" (UniqueName: \"kubernetes.io/projected/fddde3da-8512-4e62-9c38-b59f98e117e0-kube-api-access-4fw6r\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.160989 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/892d15b6-460e-4892-a836-0cc284c8a326-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mthhs\" (UID: \"892d15b6-460e-4892-a836-0cc284c8a326\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161061 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/12b1cc27-6a60-43d2-9d5e-eb7a54c1e899-trusted-ca\") pod \"console-operator-58897d9998-fc4z8\" (UID: \"12b1cc27-6a60-43d2-9d5e-eb7a54c1e899\") " pod="openshift-console-operator/console-operator-58897d9998-fc4z8" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161041 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8bb7828f-6d99-4539-8312-c8e96bfbc608-audit-policies\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161110 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h4n4\" (UniqueName: \"kubernetes.io/projected/68d61e65-8275-4862-9dae-a75029889b2a-kube-api-access-8h4n4\") pod \"machine-config-controller-84d6567774-57cxh\" (UID: \"68d61e65-8275-4862-9dae-a75029889b2a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-57cxh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161152 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1017f8be-2192-46a4-8717-ead73ca5e81b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2zh8p\" (UID: \"1017f8be-2192-46a4-8717-ead73ca5e81b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zh8p" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161169 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eddac2a3-8b07-410d-9e9f-f7ce0c05abf5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-725jh\" (UID: \"eddac2a3-8b07-410d-9e9f-f7ce0c05abf5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-725jh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161192 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/10a9b5d9-308c-4971-a073-9c88de98d8ea-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9v866\" (UID: \"10a9b5d9-308c-4971-a073-9c88de98d8ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9v866" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161231 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a79bbac6-f40a-4c92-8854-7ab5e72573cc-service-ca-bundle\") pod \"router-default-5444994796-2z46g\" (UID: \"a79bbac6-f40a-4c92-8854-7ab5e72573cc\") " pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161251 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b0b465ff-f8f3-4a99-9235-1ddc4ce093e4-srv-cert\") pod \"olm-operator-6b444d44fb-v296h\" (UID: \"b0b465ff-f8f3-4a99-9235-1ddc4ce093e4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v296h" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161305 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/edb67e99-0d50-46e9-adbb-c0831dd915d8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-xv7nw\" (UID: \"edb67e99-0d50-46e9-adbb-c0831dd915d8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xv7nw" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161335 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-console-config\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161357 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/918078e1-af38-475a-86d5-8179cafa18db-serving-cert\") pod \"service-ca-operator-777779d784-wqjnj\" (UID: \"918078e1-af38-475a-86d5-8179cafa18db\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wqjnj" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161378 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5fwg\" (UniqueName: \"kubernetes.io/projected/02b06614-31da-49d9-bc97-cf61b065d42f-kube-api-access-s5fwg\") pod \"machine-config-operator-74547568cd-8kpqj\" (UID: \"02b06614-31da-49d9-bc97-cf61b065d42f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161402 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe3df1e1-4c22-48df-aaea-469c864f0310-config\") pod \"machine-api-operator-5694c8668f-5rnj6\" (UID: \"fe3df1e1-4c22-48df-aaea-469c864f0310\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5rnj6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161428 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-td4ls\" (UniqueName: \"kubernetes.io/projected/10a9b5d9-308c-4971-a073-9c88de98d8ea-kube-api-access-td4ls\") pod \"cluster-image-registry-operator-dc59b4c8b-9v866\" (UID: \"10a9b5d9-308c-4971-a073-9c88de98d8ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9v866" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161448 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161468 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-serving-cert\") pod \"controller-manager-879f6c89f-r6c6z\" (UID: \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161485 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161500 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-audit-policies\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161516 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6e22b5c7-4191-4f21-82ba-3014ccc4e978-etcd-serving-ca\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161533 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm9fk\" (UniqueName: \"kubernetes.io/projected/a5aa6958-e573-4efb-a031-218c62b0bec9-kube-api-access-wm9fk\") pod \"collect-profiles-29567805-97bct\" (UID: \"a5aa6958-e573-4efb-a031-218c62b0bec9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161559 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-r6c6z\" (UID: \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161588 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9de83f0b-7dd2-4846-a1ce-c8af930778f4-mountpoint-dir\") pod \"csi-hostpathplugin-hvfsp\" (UID: \"9de83f0b-7dd2-4846-a1ce-c8af930778f4\") " pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161613 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161628 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6e22b5c7-4191-4f21-82ba-3014ccc4e978-encryption-config\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161648 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vd8k\" (UniqueName: \"kubernetes.io/projected/1d080332-c215-44c1-a027-65afbc612f88-kube-api-access-6vd8k\") pod \"service-ca-9c57cc56f-gd4gc\" (UID: \"1d080332-c215-44c1-a027-65afbc612f88\") " pod="openshift-service-ca/service-ca-9c57cc56f-gd4gc" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161668 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1017f8be-2192-46a4-8717-ead73ca5e81b-config\") pod \"kube-apiserver-operator-766d6c64bb-2zh8p\" (UID: \"1017f8be-2192-46a4-8717-ead73ca5e81b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zh8p" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161688 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/918078e1-af38-475a-86d5-8179cafa18db-config\") pod \"service-ca-operator-777779d784-wqjnj\" (UID: \"918078e1-af38-475a-86d5-8179cafa18db\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wqjnj" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161757 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1017f8be-2192-46a4-8717-ead73ca5e81b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2zh8p\" (UID: \"1017f8be-2192-46a4-8717-ead73ca5e81b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zh8p" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161780 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nbfq\" (UniqueName: \"kubernetes.io/projected/2d2512b8-9fb6-4b9b-9ac2-268e0b8bd04a-kube-api-access-7nbfq\") pod \"openshift-apiserver-operator-796bbdcf4f-5pdcr\" (UID: \"2d2512b8-9fb6-4b9b-9ac2-268e0b8bd04a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5pdcr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161796 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqkw4\" (UniqueName: \"kubernetes.io/projected/8bb7828f-6d99-4539-8312-c8e96bfbc608-kube-api-access-lqkw4\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161813 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-trusted-ca-bundle\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161834 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-oauth-serving-cert\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161850 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm89b\" (UniqueName: \"kubernetes.io/projected/cd25c8a4-8047-4602-a95b-3308af65bd38-kube-api-access-gm89b\") pod \"auto-csr-approver-29567810-wb89g\" (UID: \"cd25c8a4-8047-4602-a95b-3308af65bd38\") " pod="openshift-infra/auto-csr-approver-29567810-wb89g" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161870 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6e22b5c7-4191-4f21-82ba-3014ccc4e978-audit-dir\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161890 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/edb4245a-7971-4c40-81b6-27d56b319a2f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-k5dks\" (UID: \"edb4245a-7971-4c40-81b6-27d56b319a2f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k5dks" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161911 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhqjf\" (UniqueName: \"kubernetes.io/projected/a79bbac6-f40a-4c92-8854-7ab5e72573cc-kube-api-access-nhqjf\") pod \"router-default-5444994796-2z46g\" (UID: \"a79bbac6-f40a-4c92-8854-7ab5e72573cc\") " pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161930 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqpfp\" (UniqueName: \"kubernetes.io/projected/bc2b8485-8cc1-4029-8318-397d4278e455-kube-api-access-tqpfp\") pod \"package-server-manager-789f6589d5-mbgvc\" (UID: \"bc2b8485-8cc1-4029-8318-397d4278e455\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mbgvc" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161937 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/768c9343-1391-4221-b008-5dee2921953f-config\") pod \"authentication-operator-69f744f599-dzxd5\" (UID: \"768c9343-1391-4221-b008-5dee2921953f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dzxd5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161953 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.161992 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6e22b5c7-4191-4f21-82ba-3014ccc4e978-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.162025 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/892d15b6-460e-4892-a836-0cc284c8a326-serving-cert\") pod \"openshift-config-operator-7777fb866f-mthhs\" (UID: \"892d15b6-460e-4892-a836-0cc284c8a326\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.162073 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c52b4\" (UniqueName: \"kubernetes.io/projected/0c0dc3b9-8e11-4ff0-8f26-549ce47215f0-kube-api-access-c52b4\") pod \"kube-storage-version-migrator-operator-b67b599dd-blwmr\" (UID: \"0c0dc3b9-8e11-4ff0-8f26-549ce47215f0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-blwmr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.162165 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9395e24-0d4b-4165-bf60-068876927f58-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5ztsd\" (UID: \"f9395e24-0d4b-4165-bf60-068876927f58\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5ztsd" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.162250 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/892d15b6-460e-4892-a836-0cc284c8a326-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mthhs\" (UID: \"892d15b6-460e-4892-a836-0cc284c8a326\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.162300 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.162316 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8bb7828f-6d99-4539-8312-c8e96bfbc608-audit-policies\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.162389 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-service-ca\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.162418 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/10a9b5d9-308c-4971-a073-9c88de98d8ea-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9v866\" (UID: \"10a9b5d9-308c-4971-a073-9c88de98d8ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9v866" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.162445 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/07720715-40fb-4f7d-8f6b-381dc7cf9ea2-etcd-client\") pod \"etcd-operator-b45778765-4nfs6\" (UID: \"07720715-40fb-4f7d-8f6b-381dc7cf9ea2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.162509 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e5224539-6d29-4bc3-9656-4665eb287e28-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-d7xpn\" (UID: \"e5224539-6d29-4bc3-9656-4665eb287e28\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-d7xpn" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.162535 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8bb7828f-6d99-4539-8312-c8e96bfbc608-audit-dir\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.162664 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eddac2a3-8b07-410d-9e9f-f7ce0c05abf5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-725jh\" (UID: \"eddac2a3-8b07-410d-9e9f-f7ce0c05abf5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-725jh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.162701 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-blwmr"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.162688 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9de83f0b-7dd2-4846-a1ce-c8af930778f4-plugins-dir\") pod \"csi-hostpathplugin-hvfsp\" (UID: \"9de83f0b-7dd2-4846-a1ce-c8af930778f4\") " pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.162792 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/768c9343-1391-4221-b008-5dee2921953f-service-ca-bundle\") pod \"authentication-operator-69f744f599-dzxd5\" (UID: \"768c9343-1391-4221-b008-5dee2921953f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dzxd5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.162826 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fddde3da-8512-4e62-9c38-b59f98e117e0-audit-dir\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.162888 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae87529c-e52f-45b0-9bbf-2a652e628bc5-config-volume\") pod \"dns-default-zs5pr\" (UID: \"ae87529c-e52f-45b0-9bbf-2a652e628bc5\") " pod="openshift-dns/dns-default-zs5pr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.162906 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.162921 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9395e24-0d4b-4165-bf60-068876927f58-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5ztsd\" (UID: \"f9395e24-0d4b-4165-bf60-068876927f58\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5ztsd" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.162983 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/10a9b5d9-308c-4971-a073-9c88de98d8ea-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9v866\" (UID: \"10a9b5d9-308c-4971-a073-9c88de98d8ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9v866" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.163013 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6343dd4f-5c5b-4c94-a4d6-f603698ba6ac-machine-approver-tls\") pod \"machine-approver-56656f9798-468g4\" (UID: \"6343dd4f-5c5b-4c94-a4d6-f603698ba6ac\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-468g4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.163046 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8bb7828f-6d99-4539-8312-c8e96bfbc608-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.163077 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9de83f0b-7dd2-4846-a1ce-c8af930778f4-socket-dir\") pod \"csi-hostpathplugin-hvfsp\" (UID: \"9de83f0b-7dd2-4846-a1ce-c8af930778f4\") " pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.163221 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9blnw\" (UniqueName: \"kubernetes.io/projected/a49709c7-59e0-440e-89c2-177c42cd28e8-kube-api-access-9blnw\") pod \"packageserver-d55dfcdfc-sp69c\" (UID: \"a49709c7-59e0-440e-89c2-177c42cd28e8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.163254 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-client-ca\") pod \"controller-manager-879f6c89f-r6c6z\" (UID: \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.163285 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8bb7828f-6d99-4539-8312-c8e96bfbc608-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.163314 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c0dc3b9-8e11-4ff0-8f26-549ce47215f0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-blwmr\" (UID: \"0c0dc3b9-8e11-4ff0-8f26-549ce47215f0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-blwmr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.163339 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8bb7828f-6d99-4539-8312-c8e96bfbc608-serving-cert\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.163369 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvbpw\" (UniqueName: \"kubernetes.io/projected/edb4245a-7971-4c40-81b6-27d56b319a2f-kube-api-access-dvbpw\") pod \"cluster-samples-operator-665b6dd947-k5dks\" (UID: \"edb4245a-7971-4c40-81b6-27d56b319a2f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k5dks" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.163427 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a79bbac6-f40a-4c92-8854-7ab5e72573cc-metrics-certs\") pod \"router-default-5444994796-2z46g\" (UID: \"a79bbac6-f40a-4c92-8854-7ab5e72573cc\") " pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.163429 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.163550 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.163628 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6343dd4f-5c5b-4c94-a4d6-f603698ba6ac-config\") pod \"machine-approver-56656f9798-468g4\" (UID: \"6343dd4f-5c5b-4c94-a4d6-f603698ba6ac\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-468g4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.163655 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.163684 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/02b06614-31da-49d9-bc97-cf61b065d42f-images\") pod \"machine-config-operator-74547568cd-8kpqj\" (UID: \"02b06614-31da-49d9-bc97-cf61b065d42f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.163732 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1d080332-c215-44c1-a027-65afbc612f88-signing-key\") pod \"service-ca-9c57cc56f-gd4gc\" (UID: \"1d080332-c215-44c1-a027-65afbc612f88\") " pod="openshift-service-ca/service-ca-9c57cc56f-gd4gc" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.163759 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5aa6958-e573-4efb-a031-218c62b0bec9-secret-volume\") pod \"collect-profiles-29567805-97bct\" (UID: \"a5aa6958-e573-4efb-a031-218c62b0bec9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.163820 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m6fm\" (UniqueName: \"kubernetes.io/projected/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-kube-api-access-8m6fm\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.163927 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a49709c7-59e0-440e-89c2-177c42cd28e8-apiservice-cert\") pod \"packageserver-d55dfcdfc-sp69c\" (UID: \"a49709c7-59e0-440e-89c2-177c42cd28e8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.163989 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe3df1e1-4c22-48df-aaea-469c864f0310-config\") pod \"machine-api-operator-5694c8668f-5rnj6\" (UID: \"fe3df1e1-4c22-48df-aaea-469c864f0310\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5rnj6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.164006 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07720715-40fb-4f7d-8f6b-381dc7cf9ea2-config\") pod \"etcd-operator-b45778765-4nfs6\" (UID: \"07720715-40fb-4f7d-8f6b-381dc7cf9ea2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.164205 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1d080332-c215-44c1-a027-65afbc612f88-signing-cabundle\") pod \"service-ca-9c57cc56f-gd4gc\" (UID: \"1d080332-c215-44c1-a027-65afbc612f88\") " pod="openshift-service-ca/service-ca-9c57cc56f-gd4gc" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.164285 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8bb7828f-6d99-4539-8312-c8e96bfbc608-etcd-client\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.164309 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d2512b8-9fb6-4b9b-9ac2-268e0b8bd04a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5pdcr\" (UID: \"2d2512b8-9fb6-4b9b-9ac2-268e0b8bd04a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5pdcr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.164339 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/07720715-40fb-4f7d-8f6b-381dc7cf9ea2-etcd-ca\") pod \"etcd-operator-b45778765-4nfs6\" (UID: \"07720715-40fb-4f7d-8f6b-381dc7cf9ea2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.164358 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d2512b8-9fb6-4b9b-9ac2-268e0b8bd04a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5pdcr\" (UID: \"2d2512b8-9fb6-4b9b-9ac2-268e0b8bd04a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5pdcr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.164386 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7h98\" (UniqueName: \"kubernetes.io/projected/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-kube-api-access-r7h98\") pod \"route-controller-manager-6576b87f9c-55lcp\" (UID: \"ed9baf4c-70b2-450f-9b21-76dfafbc44d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.164392 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-service-ca\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.164406 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk47r\" (UniqueName: \"kubernetes.io/projected/12b1cc27-6a60-43d2-9d5e-eb7a54c1e899-kube-api-access-rk47r\") pod \"console-operator-58897d9998-fc4z8\" (UID: \"12b1cc27-6a60-43d2-9d5e-eb7a54c1e899\") " pod="openshift-console-operator/console-operator-58897d9998-fc4z8" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.164425 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edb67e99-0d50-46e9-adbb-c0831dd915d8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-xv7nw\" (UID: \"edb67e99-0d50-46e9-adbb-c0831dd915d8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xv7nw" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.164449 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/bc2b8485-8cc1-4029-8318-397d4278e455-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-mbgvc\" (UID: \"bc2b8485-8cc1-4029-8318-397d4278e455\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mbgvc" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.164489 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9de83f0b-7dd2-4846-a1ce-c8af930778f4-registration-dir\") pod \"csi-hostpathplugin-hvfsp\" (UID: \"9de83f0b-7dd2-4846-a1ce-c8af930778f4\") " pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.164535 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8bb7828f-6d99-4539-8312-c8e96bfbc608-encryption-config\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.164559 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6e22b5c7-4191-4f21-82ba-3014ccc4e978-image-import-ca\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.164595 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-console-oauth-config\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.164069 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/768c9343-1391-4221-b008-5dee2921953f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dzxd5\" (UID: \"768c9343-1391-4221-b008-5dee2921953f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dzxd5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.165444 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6e22b5c7-4191-4f21-82ba-3014ccc4e978-audit-dir\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.165410 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-r6c6z\" (UID: \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.164560 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8bb7828f-6d99-4539-8312-c8e96bfbc608-audit-dir\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.166916 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-config\") pod \"route-controller-manager-6576b87f9c-55lcp\" (UID: \"ed9baf4c-70b2-450f-9b21-76dfafbc44d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.166993 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6e22b5c7-4191-4f21-82ba-3014ccc4e978-audit\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.167076 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5ztsd"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.167087 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fe3df1e1-4c22-48df-aaea-469c864f0310-images\") pod \"machine-api-operator-5694c8668f-5rnj6\" (UID: \"fe3df1e1-4c22-48df-aaea-469c864f0310\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5rnj6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.168942 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fddde3da-8512-4e62-9c38-b59f98e117e0-audit-dir\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.169009 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6343dd4f-5c5b-4c94-a4d6-f603698ba6ac-auth-proxy-config\") pod \"machine-approver-56656f9798-468g4\" (UID: \"6343dd4f-5c5b-4c94-a4d6-f603698ba6ac\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-468g4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.169081 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/768c9343-1391-4221-b008-5dee2921953f-serving-cert\") pod \"authentication-operator-69f744f599-dzxd5\" (UID: \"768c9343-1391-4221-b008-5dee2921953f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dzxd5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.169108 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/07720715-40fb-4f7d-8f6b-381dc7cf9ea2-etcd-service-ca\") pod \"etcd-operator-b45778765-4nfs6\" (UID: \"07720715-40fb-4f7d-8f6b-381dc7cf9ea2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.169156 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r82zp\" (UniqueName: \"kubernetes.io/projected/918078e1-af38-475a-86d5-8179cafa18db-kube-api-access-r82zp\") pod \"service-ca-operator-777779d784-wqjnj\" (UID: \"918078e1-af38-475a-86d5-8179cafa18db\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wqjnj" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.169327 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d2512b8-9fb6-4b9b-9ac2-268e0b8bd04a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5pdcr\" (UID: \"2d2512b8-9fb6-4b9b-9ac2-268e0b8bd04a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5pdcr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.169709 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-oauth-serving-cert\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.170335 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.170347 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6e22b5c7-4191-4f21-82ba-3014ccc4e978-image-import-ca\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.171058 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/768c9343-1391-4221-b008-5dee2921953f-service-ca-bundle\") pod \"authentication-operator-69f744f599-dzxd5\" (UID: \"768c9343-1391-4221-b008-5dee2921953f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dzxd5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.171297 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d2512b8-9fb6-4b9b-9ac2-268e0b8bd04a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5pdcr\" (UID: \"2d2512b8-9fb6-4b9b-9ac2-268e0b8bd04a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5pdcr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.172143 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8bb7828f-6d99-4539-8312-c8e96bfbc608-etcd-client\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.172391 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.173534 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/892d15b6-460e-4892-a836-0cc284c8a326-serving-cert\") pod \"openshift-config-operator-7777fb866f-mthhs\" (UID: \"892d15b6-460e-4892-a836-0cc284c8a326\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.173576 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/edb4245a-7971-4c40-81b6-27d56b319a2f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-k5dks\" (UID: \"edb4245a-7971-4c40-81b6-27d56b319a2f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k5dks" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.173755 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.174225 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e22b5c7-4191-4f21-82ba-3014ccc4e978-serving-cert\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.174684 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-client-ca\") pod \"controller-manager-879f6c89f-r6c6z\" (UID: \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.174929 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8bb7828f-6d99-4539-8312-c8e96bfbc608-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.175739 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-serving-cert\") pod \"route-controller-manager-6576b87f9c-55lcp\" (UID: \"ed9baf4c-70b2-450f-9b21-76dfafbc44d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.175807 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/10a9b5d9-308c-4971-a073-9c88de98d8ea-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9v866\" (UID: \"10a9b5d9-308c-4971-a073-9c88de98d8ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9v866" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.176101 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/10a9b5d9-308c-4971-a073-9c88de98d8ea-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9v866\" (UID: \"10a9b5d9-308c-4971-a073-9c88de98d8ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9v866" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.176598 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8bb7828f-6d99-4539-8312-c8e96bfbc608-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.176653 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6343dd4f-5c5b-4c94-a4d6-f603698ba6ac-config\") pod \"machine-approver-56656f9798-468g4\" (UID: \"6343dd4f-5c5b-4c94-a4d6-f603698ba6ac\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-468g4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.176980 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-trusted-ca-bundle\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.177330 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.178738 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.178840 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-console-oauth-config\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.179000 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6343dd4f-5c5b-4c94-a4d6-f603698ba6ac-machine-approver-tls\") pod \"machine-approver-56656f9798-468g4\" (UID: \"6343dd4f-5c5b-4c94-a4d6-f603698ba6ac\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-468g4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.179223 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6e22b5c7-4191-4f21-82ba-3014ccc4e978-audit\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.179672 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-audit-policies\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.179739 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12b1cc27-6a60-43d2-9d5e-eb7a54c1e899-serving-cert\") pod \"console-operator-58897d9998-fc4z8\" (UID: \"12b1cc27-6a60-43d2-9d5e-eb7a54c1e899\") " pod="openshift-console-operator/console-operator-58897d9998-fc4z8" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.179846 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-console-config\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.180058 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6e22b5c7-4191-4f21-82ba-3014ccc4e978-etcd-serving-ca\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.180098 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8bb7828f-6d99-4539-8312-c8e96bfbc608-serving-cert\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.180291 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-config\") pod \"route-controller-manager-6576b87f9c-55lcp\" (UID: \"ed9baf4c-70b2-450f-9b21-76dfafbc44d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.180526 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6343dd4f-5c5b-4c94-a4d6-f603698ba6ac-auth-proxy-config\") pod \"machine-approver-56656f9798-468g4\" (UID: \"6343dd4f-5c5b-4c94-a4d6-f603698ba6ac\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-468g4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.180899 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.180918 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fe3df1e1-4c22-48df-aaea-469c864f0310-images\") pod \"machine-api-operator-5694c8668f-5rnj6\" (UID: \"fe3df1e1-4c22-48df-aaea-469c864f0310\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5rnj6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.169228 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.181034 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.181153 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m44cw\" (UniqueName: \"kubernetes.io/projected/dfbaac71-f99c-4373-a469-f2e5dd0ee632-kube-api-access-m44cw\") pod \"marketplace-operator-79b997595-wrzs4\" (UID: \"dfbaac71-f99c-4373-a469-f2e5dd0ee632\") " pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.182078 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.182226 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.182252 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75ck2\" (UniqueName: \"kubernetes.io/projected/6e22b5c7-4191-4f21-82ba-3014ccc4e978-kube-api-access-75ck2\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.182362 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td8g7\" (UniqueName: \"kubernetes.io/projected/ae87529c-e52f-45b0-9bbf-2a652e628bc5-kube-api-access-td8g7\") pod \"dns-default-zs5pr\" (UID: \"ae87529c-e52f-45b0-9bbf-2a652e628bc5\") " pod="openshift-dns/dns-default-zs5pr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.182697 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpvcx\" (UniqueName: \"kubernetes.io/projected/811f064c-ebf4-48ad-87a0-83205eb1eca5-kube-api-access-kpvcx\") pod \"dns-operator-744455d44c-vjv4q\" (UID: \"811f064c-ebf4-48ad-87a0-83205eb1eca5\") " pod="openshift-dns-operator/dns-operator-744455d44c-vjv4q" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.182744 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6e22b5c7-4191-4f21-82ba-3014ccc4e978-etcd-client\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.182924 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12b1cc27-6a60-43d2-9d5e-eb7a54c1e899-config\") pod \"console-operator-58897d9998-fc4z8\" (UID: \"12b1cc27-6a60-43d2-9d5e-eb7a54c1e899\") " pod="openshift-console-operator/console-operator-58897d9998-fc4z8" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.182944 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.183027 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/fe3df1e1-4c22-48df-aaea-469c864f0310-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5rnj6\" (UID: \"fe3df1e1-4c22-48df-aaea-469c864f0310\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5rnj6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.183278 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/02b06614-31da-49d9-bc97-cf61b065d42f-proxy-tls\") pod \"machine-config-operator-74547568cd-8kpqj\" (UID: \"02b06614-31da-49d9-bc97-cf61b065d42f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.183292 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8bb7828f-6d99-4539-8312-c8e96bfbc608-encryption-config\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.183314 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b0b465ff-f8f3-4a99-9235-1ddc4ce093e4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-v296h\" (UID: \"b0b465ff-f8f3-4a99-9235-1ddc4ce093e4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v296h" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.183342 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9de83f0b-7dd2-4846-a1ce-c8af930778f4-csi-data-dir\") pod \"csi-hostpathplugin-hvfsp\" (UID: \"9de83f0b-7dd2-4846-a1ce-c8af930778f4\") " pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.183069 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/768c9343-1391-4221-b008-5dee2921953f-serving-cert\") pod \"authentication-operator-69f744f599-dzxd5\" (UID: \"768c9343-1391-4221-b008-5dee2921953f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dzxd5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.183475 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e22b5c7-4191-4f21-82ba-3014ccc4e978-config\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.183586 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/68d61e65-8275-4862-9dae-a75029889b2a-proxy-tls\") pod \"machine-config-controller-84d6567774-57cxh\" (UID: \"68d61e65-8275-4862-9dae-a75029889b2a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-57cxh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.183621 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/02b06614-31da-49d9-bc97-cf61b065d42f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8kpqj\" (UID: \"02b06614-31da-49d9-bc97-cf61b065d42f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.183752 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-client-ca\") pod \"route-controller-manager-6576b87f9c-55lcp\" (UID: \"ed9baf4c-70b2-450f-9b21-76dfafbc44d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.183798 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dfbaac71-f99c-4373-a469-f2e5dd0ee632-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wrzs4\" (UID: \"dfbaac71-f99c-4373-a469-f2e5dd0ee632\") " pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.184009 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5aa6958-e573-4efb-a031-218c62b0bec9-config-volume\") pod \"collect-profiles-29567805-97bct\" (UID: \"a5aa6958-e573-4efb-a031-218c62b0bec9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.184053 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12b1cc27-6a60-43d2-9d5e-eb7a54c1e899-config\") pod \"console-operator-58897d9998-fc4z8\" (UID: \"12b1cc27-6a60-43d2-9d5e-eb7a54c1e899\") " pod="openshift-console-operator/console-operator-58897d9998-fc4z8" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.184087 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-config\") pod \"controller-manager-879f6c89f-r6c6z\" (UID: \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.184831 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e22b5c7-4191-4f21-82ba-3014ccc4e978-config\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.185237 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-client-ca\") pod \"route-controller-manager-6576b87f9c-55lcp\" (UID: \"ed9baf4c-70b2-450f-9b21-76dfafbc44d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.185798 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rz6g5"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.186637 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-config\") pod \"controller-manager-879f6c89f-r6c6z\" (UID: \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.187831 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6e22b5c7-4191-4f21-82ba-3014ccc4e978-encryption-config\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.188473 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/fe3df1e1-4c22-48df-aaea-469c864f0310-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5rnj6\" (UID: \"fe3df1e1-4c22-48df-aaea-469c864f0310\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5rnj6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.188697 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.188835 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-serving-cert\") pod \"controller-manager-879f6c89f-r6c6z\" (UID: \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.190542 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6e22b5c7-4191-4f21-82ba-3014ccc4e978-etcd-client\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.190603 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.190794 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-console-serving-cert\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.192535 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wrzs4"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.194562 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2lnkp"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.196280 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wqjnj"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.197439 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567810-wb89g"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.198533 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-725jh"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.199531 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.200177 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.200549 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.201627 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zh8p"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.202832 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-6jnn5"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.203772 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-6jnn5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.203885 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-dnv7h"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.205646 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-6jnn5"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.205798 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-dnv7h" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.206382 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.207611 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xv7nw"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.208700 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-hvfsp"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.209743 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-zs5pr"] Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.220504 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.236927 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"02f610dfd36fe9bba151d035cc95b083bc2a0a134823b67b5fe8dd4a65fb82dd"} Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.236974 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c99c97d573d15e53e325a18910ea38150805d0399c3dc90e8de6aa020d44a415"} Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.237165 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.238365 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"1a8243a3daa4a9d245ddf7c678db9d2493ed45293514f40f1e2db64d4e750ffb"} Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.238410 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d34204adb9c00482355cdebd845da5821eccc5e49d6d52361ac87a4148502dfd"} Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.239870 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.260303 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.280332 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285211 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1d080332-c215-44c1-a027-65afbc612f88-signing-cabundle\") pod \"service-ca-9c57cc56f-gd4gc\" (UID: \"1d080332-c215-44c1-a027-65afbc612f88\") " pod="openshift-service-ca/service-ca-9c57cc56f-gd4gc" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285249 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/07720715-40fb-4f7d-8f6b-381dc7cf9ea2-etcd-ca\") pod \"etcd-operator-b45778765-4nfs6\" (UID: \"07720715-40fb-4f7d-8f6b-381dc7cf9ea2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285282 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/bc2b8485-8cc1-4029-8318-397d4278e455-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-mbgvc\" (UID: \"bc2b8485-8cc1-4029-8318-397d4278e455\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mbgvc" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285300 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edb67e99-0d50-46e9-adbb-c0831dd915d8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-xv7nw\" (UID: \"edb67e99-0d50-46e9-adbb-c0831dd915d8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xv7nw" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285317 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9de83f0b-7dd2-4846-a1ce-c8af930778f4-registration-dir\") pod \"csi-hostpathplugin-hvfsp\" (UID: \"9de83f0b-7dd2-4846-a1ce-c8af930778f4\") " pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285337 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/07720715-40fb-4f7d-8f6b-381dc7cf9ea2-etcd-service-ca\") pod \"etcd-operator-b45778765-4nfs6\" (UID: \"07720715-40fb-4f7d-8f6b-381dc7cf9ea2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285354 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m44cw\" (UniqueName: \"kubernetes.io/projected/dfbaac71-f99c-4373-a469-f2e5dd0ee632-kube-api-access-m44cw\") pod \"marketplace-operator-79b997595-wrzs4\" (UID: \"dfbaac71-f99c-4373-a469-f2e5dd0ee632\") " pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285372 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r82zp\" (UniqueName: \"kubernetes.io/projected/918078e1-af38-475a-86d5-8179cafa18db-kube-api-access-r82zp\") pod \"service-ca-operator-777779d784-wqjnj\" (UID: \"918078e1-af38-475a-86d5-8179cafa18db\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wqjnj" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285397 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-td8g7\" (UniqueName: \"kubernetes.io/projected/ae87529c-e52f-45b0-9bbf-2a652e628bc5-kube-api-access-td8g7\") pod \"dns-default-zs5pr\" (UID: \"ae87529c-e52f-45b0-9bbf-2a652e628bc5\") " pod="openshift-dns/dns-default-zs5pr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285415 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpvcx\" (UniqueName: \"kubernetes.io/projected/811f064c-ebf4-48ad-87a0-83205eb1eca5-kube-api-access-kpvcx\") pod \"dns-operator-744455d44c-vjv4q\" (UID: \"811f064c-ebf4-48ad-87a0-83205eb1eca5\") " pod="openshift-dns-operator/dns-operator-744455d44c-vjv4q" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285434 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/02b06614-31da-49d9-bc97-cf61b065d42f-proxy-tls\") pod \"machine-config-operator-74547568cd-8kpqj\" (UID: \"02b06614-31da-49d9-bc97-cf61b065d42f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285484 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/68d61e65-8275-4862-9dae-a75029889b2a-proxy-tls\") pod \"machine-config-controller-84d6567774-57cxh\" (UID: \"68d61e65-8275-4862-9dae-a75029889b2a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-57cxh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285503 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/02b06614-31da-49d9-bc97-cf61b065d42f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8kpqj\" (UID: \"02b06614-31da-49d9-bc97-cf61b065d42f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285522 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b0b465ff-f8f3-4a99-9235-1ddc4ce093e4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-v296h\" (UID: \"b0b465ff-f8f3-4a99-9235-1ddc4ce093e4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v296h" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285568 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9de83f0b-7dd2-4846-a1ce-c8af930778f4-csi-data-dir\") pod \"csi-hostpathplugin-hvfsp\" (UID: \"9de83f0b-7dd2-4846-a1ce-c8af930778f4\") " pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285587 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dfbaac71-f99c-4373-a469-f2e5dd0ee632-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wrzs4\" (UID: \"dfbaac71-f99c-4373-a469-f2e5dd0ee632\") " pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285616 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5aa6958-e573-4efb-a031-218c62b0bec9-config-volume\") pod \"collect-profiles-29567805-97bct\" (UID: \"a5aa6958-e573-4efb-a031-218c62b0bec9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285632 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dfbaac71-f99c-4373-a469-f2e5dd0ee632-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wrzs4\" (UID: \"dfbaac71-f99c-4373-a469-f2e5dd0ee632\") " pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285650 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f9395e24-0d4b-4165-bf60-068876927f58-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5ztsd\" (UID: \"f9395e24-0d4b-4165-bf60-068876927f58\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5ztsd" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285670 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9de83f0b-7dd2-4846-a1ce-c8af930778f4-registration-dir\") pod \"csi-hostpathplugin-hvfsp\" (UID: \"9de83f0b-7dd2-4846-a1ce-c8af930778f4\") " pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285699 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9de83f0b-7dd2-4846-a1ce-c8af930778f4-csi-data-dir\") pod \"csi-hostpathplugin-hvfsp\" (UID: \"9de83f0b-7dd2-4846-a1ce-c8af930778f4\") " pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285677 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07720715-40fb-4f7d-8f6b-381dc7cf9ea2-serving-cert\") pod \"etcd-operator-b45778765-4nfs6\" (UID: \"07720715-40fb-4f7d-8f6b-381dc7cf9ea2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285885 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clmnm\" (UniqueName: \"kubernetes.io/projected/9de83f0b-7dd2-4846-a1ce-c8af930778f4-kube-api-access-clmnm\") pod \"csi-hostpathplugin-hvfsp\" (UID: \"9de83f0b-7dd2-4846-a1ce-c8af930778f4\") " pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285912 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae87529c-e52f-45b0-9bbf-2a652e628bc5-metrics-tls\") pod \"dns-default-zs5pr\" (UID: \"ae87529c-e52f-45b0-9bbf-2a652e628bc5\") " pod="openshift-dns/dns-default-zs5pr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285936 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eddac2a3-8b07-410d-9e9f-f7ce0c05abf5-config\") pod \"kube-controller-manager-operator-78b949d7b-725jh\" (UID: \"eddac2a3-8b07-410d-9e9f-f7ce0c05abf5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-725jh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285955 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c0dc3b9-8e11-4ff0-8f26-549ce47215f0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-blwmr\" (UID: \"0c0dc3b9-8e11-4ff0-8f26-549ce47215f0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-blwmr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285974 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9mb9\" (UniqueName: \"kubernetes.io/projected/edb67e99-0d50-46e9-adbb-c0831dd915d8-kube-api-access-d9mb9\") pod \"openshift-controller-manager-operator-756b6f6bc6-xv7nw\" (UID: \"edb67e99-0d50-46e9-adbb-c0831dd915d8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xv7nw" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.285997 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a79bbac6-f40a-4c92-8854-7ab5e72573cc-stats-auth\") pod \"router-default-5444994796-2z46g\" (UID: \"a79bbac6-f40a-4c92-8854-7ab5e72573cc\") " pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286014 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nw58z\" (UniqueName: \"kubernetes.io/projected/07720715-40fb-4f7d-8f6b-381dc7cf9ea2-kube-api-access-nw58z\") pod \"etcd-operator-b45778765-4nfs6\" (UID: \"07720715-40fb-4f7d-8f6b-381dc7cf9ea2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286035 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a49709c7-59e0-440e-89c2-177c42cd28e8-tmpfs\") pod \"packageserver-d55dfcdfc-sp69c\" (UID: \"a49709c7-59e0-440e-89c2-177c42cd28e8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286052 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zfpw\" (UniqueName: \"kubernetes.io/projected/e5224539-6d29-4bc3-9656-4665eb287e28-kube-api-access-7zfpw\") pod \"control-plane-machine-set-operator-78cbb6b69f-d7xpn\" (UID: \"e5224539-6d29-4bc3-9656-4665eb287e28\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-d7xpn" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286146 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plx8q\" (UniqueName: \"kubernetes.io/projected/b0b465ff-f8f3-4a99-9235-1ddc4ce093e4-kube-api-access-plx8q\") pod \"olm-operator-6b444d44fb-v296h\" (UID: \"b0b465ff-f8f3-4a99-9235-1ddc4ce093e4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v296h" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286165 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a49709c7-59e0-440e-89c2-177c42cd28e8-webhook-cert\") pod \"packageserver-d55dfcdfc-sp69c\" (UID: \"a49709c7-59e0-440e-89c2-177c42cd28e8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286183 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/68d61e65-8275-4862-9dae-a75029889b2a-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-57cxh\" (UID: \"68d61e65-8275-4862-9dae-a75029889b2a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-57cxh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286198 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/811f064c-ebf4-48ad-87a0-83205eb1eca5-metrics-tls\") pod \"dns-operator-744455d44c-vjv4q\" (UID: \"811f064c-ebf4-48ad-87a0-83205eb1eca5\") " pod="openshift-dns-operator/dns-operator-744455d44c-vjv4q" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286230 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gr9p\" (UniqueName: \"kubernetes.io/projected/2dbc150e-3a25-4b44-b01c-effe99de5152-kube-api-access-5gr9p\") pod \"migrator-59844c95c7-7sfkh\" (UID: \"2dbc150e-3a25-4b44-b01c-effe99de5152\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7sfkh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286246 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a79bbac6-f40a-4c92-8854-7ab5e72573cc-default-certificate\") pod \"router-default-5444994796-2z46g\" (UID: \"a79bbac6-f40a-4c92-8854-7ab5e72573cc\") " pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286270 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eddac2a3-8b07-410d-9e9f-f7ce0c05abf5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-725jh\" (UID: \"eddac2a3-8b07-410d-9e9f-f7ce0c05abf5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-725jh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286286 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h4n4\" (UniqueName: \"kubernetes.io/projected/68d61e65-8275-4862-9dae-a75029889b2a-kube-api-access-8h4n4\") pod \"machine-config-controller-84d6567774-57cxh\" (UID: \"68d61e65-8275-4862-9dae-a75029889b2a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-57cxh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286306 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1017f8be-2192-46a4-8717-ead73ca5e81b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2zh8p\" (UID: \"1017f8be-2192-46a4-8717-ead73ca5e81b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zh8p" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286323 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a79bbac6-f40a-4c92-8854-7ab5e72573cc-service-ca-bundle\") pod \"router-default-5444994796-2z46g\" (UID: \"a79bbac6-f40a-4c92-8854-7ab5e72573cc\") " pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286339 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b0b465ff-f8f3-4a99-9235-1ddc4ce093e4-srv-cert\") pod \"olm-operator-6b444d44fb-v296h\" (UID: \"b0b465ff-f8f3-4a99-9235-1ddc4ce093e4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v296h" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286341 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/02b06614-31da-49d9-bc97-cf61b065d42f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8kpqj\" (UID: \"02b06614-31da-49d9-bc97-cf61b065d42f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286356 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/edb67e99-0d50-46e9-adbb-c0831dd915d8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-xv7nw\" (UID: \"edb67e99-0d50-46e9-adbb-c0831dd915d8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xv7nw" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286374 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/918078e1-af38-475a-86d5-8179cafa18db-serving-cert\") pod \"service-ca-operator-777779d784-wqjnj\" (UID: \"918078e1-af38-475a-86d5-8179cafa18db\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wqjnj" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286396 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5fwg\" (UniqueName: \"kubernetes.io/projected/02b06614-31da-49d9-bc97-cf61b065d42f-kube-api-access-s5fwg\") pod \"machine-config-operator-74547568cd-8kpqj\" (UID: \"02b06614-31da-49d9-bc97-cf61b065d42f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286413 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm9fk\" (UniqueName: \"kubernetes.io/projected/a5aa6958-e573-4efb-a031-218c62b0bec9-kube-api-access-wm9fk\") pod \"collect-profiles-29567805-97bct\" (UID: \"a5aa6958-e573-4efb-a031-218c62b0bec9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286428 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9de83f0b-7dd2-4846-a1ce-c8af930778f4-mountpoint-dir\") pod \"csi-hostpathplugin-hvfsp\" (UID: \"9de83f0b-7dd2-4846-a1ce-c8af930778f4\") " pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286444 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vd8k\" (UniqueName: \"kubernetes.io/projected/1d080332-c215-44c1-a027-65afbc612f88-kube-api-access-6vd8k\") pod \"service-ca-9c57cc56f-gd4gc\" (UID: \"1d080332-c215-44c1-a027-65afbc612f88\") " pod="openshift-service-ca/service-ca-9c57cc56f-gd4gc" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286460 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1017f8be-2192-46a4-8717-ead73ca5e81b-config\") pod \"kube-apiserver-operator-766d6c64bb-2zh8p\" (UID: \"1017f8be-2192-46a4-8717-ead73ca5e81b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zh8p" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286476 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/918078e1-af38-475a-86d5-8179cafa18db-config\") pod \"service-ca-operator-777779d784-wqjnj\" (UID: \"918078e1-af38-475a-86d5-8179cafa18db\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wqjnj" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286505 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1017f8be-2192-46a4-8717-ead73ca5e81b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2zh8p\" (UID: \"1017f8be-2192-46a4-8717-ead73ca5e81b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zh8p" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286542 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhqjf\" (UniqueName: \"kubernetes.io/projected/a79bbac6-f40a-4c92-8854-7ab5e72573cc-kube-api-access-nhqjf\") pod \"router-default-5444994796-2z46g\" (UID: \"a79bbac6-f40a-4c92-8854-7ab5e72573cc\") " pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286560 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm89b\" (UniqueName: \"kubernetes.io/projected/cd25c8a4-8047-4602-a95b-3308af65bd38-kube-api-access-gm89b\") pod \"auto-csr-approver-29567810-wb89g\" (UID: \"cd25c8a4-8047-4602-a95b-3308af65bd38\") " pod="openshift-infra/auto-csr-approver-29567810-wb89g" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286579 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqpfp\" (UniqueName: \"kubernetes.io/projected/bc2b8485-8cc1-4029-8318-397d4278e455-kube-api-access-tqpfp\") pod \"package-server-manager-789f6589d5-mbgvc\" (UID: \"bc2b8485-8cc1-4029-8318-397d4278e455\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mbgvc" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286597 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c52b4\" (UniqueName: \"kubernetes.io/projected/0c0dc3b9-8e11-4ff0-8f26-549ce47215f0-kube-api-access-c52b4\") pod \"kube-storage-version-migrator-operator-b67b599dd-blwmr\" (UID: \"0c0dc3b9-8e11-4ff0-8f26-549ce47215f0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-blwmr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286613 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9395e24-0d4b-4165-bf60-068876927f58-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5ztsd\" (UID: \"f9395e24-0d4b-4165-bf60-068876927f58\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5ztsd" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286636 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/07720715-40fb-4f7d-8f6b-381dc7cf9ea2-etcd-client\") pod \"etcd-operator-b45778765-4nfs6\" (UID: \"07720715-40fb-4f7d-8f6b-381dc7cf9ea2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286654 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e5224539-6d29-4bc3-9656-4665eb287e28-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-d7xpn\" (UID: \"e5224539-6d29-4bc3-9656-4665eb287e28\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-d7xpn" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286686 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eddac2a3-8b07-410d-9e9f-f7ce0c05abf5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-725jh\" (UID: \"eddac2a3-8b07-410d-9e9f-f7ce0c05abf5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-725jh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286725 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9de83f0b-7dd2-4846-a1ce-c8af930778f4-plugins-dir\") pod \"csi-hostpathplugin-hvfsp\" (UID: \"9de83f0b-7dd2-4846-a1ce-c8af930778f4\") " pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286752 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae87529c-e52f-45b0-9bbf-2a652e628bc5-config-volume\") pod \"dns-default-zs5pr\" (UID: \"ae87529c-e52f-45b0-9bbf-2a652e628bc5\") " pod="openshift-dns/dns-default-zs5pr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286868 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9de83f0b-7dd2-4846-a1ce-c8af930778f4-mountpoint-dir\") pod \"csi-hostpathplugin-hvfsp\" (UID: \"9de83f0b-7dd2-4846-a1ce-c8af930778f4\") " pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286948 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a49709c7-59e0-440e-89c2-177c42cd28e8-tmpfs\") pod \"packageserver-d55dfcdfc-sp69c\" (UID: \"a49709c7-59e0-440e-89c2-177c42cd28e8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.287259 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/68d61e65-8275-4862-9dae-a75029889b2a-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-57cxh\" (UID: \"68d61e65-8275-4862-9dae-a75029889b2a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-57cxh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.287416 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9395e24-0d4b-4165-bf60-068876927f58-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5ztsd\" (UID: \"f9395e24-0d4b-4165-bf60-068876927f58\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5ztsd" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.287439 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9blnw\" (UniqueName: \"kubernetes.io/projected/a49709c7-59e0-440e-89c2-177c42cd28e8-kube-api-access-9blnw\") pod \"packageserver-d55dfcdfc-sp69c\" (UID: \"a49709c7-59e0-440e-89c2-177c42cd28e8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.287459 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9de83f0b-7dd2-4846-a1ce-c8af930778f4-socket-dir\") pod \"csi-hostpathplugin-hvfsp\" (UID: \"9de83f0b-7dd2-4846-a1ce-c8af930778f4\") " pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.287479 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c0dc3b9-8e11-4ff0-8f26-549ce47215f0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-blwmr\" (UID: \"0c0dc3b9-8e11-4ff0-8f26-549ce47215f0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-blwmr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.287498 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a79bbac6-f40a-4c92-8854-7ab5e72573cc-metrics-certs\") pod \"router-default-5444994796-2z46g\" (UID: \"a79bbac6-f40a-4c92-8854-7ab5e72573cc\") " pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.287416 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9de83f0b-7dd2-4846-a1ce-c8af930778f4-plugins-dir\") pod \"csi-hostpathplugin-hvfsp\" (UID: \"9de83f0b-7dd2-4846-a1ce-c8af930778f4\") " pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.286749 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eddac2a3-8b07-410d-9e9f-f7ce0c05abf5-config\") pod \"kube-controller-manager-operator-78b949d7b-725jh\" (UID: \"eddac2a3-8b07-410d-9e9f-f7ce0c05abf5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-725jh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.287522 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/02b06614-31da-49d9-bc97-cf61b065d42f-images\") pod \"machine-config-operator-74547568cd-8kpqj\" (UID: \"02b06614-31da-49d9-bc97-cf61b065d42f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.287598 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9de83f0b-7dd2-4846-a1ce-c8af930778f4-socket-dir\") pod \"csi-hostpathplugin-hvfsp\" (UID: \"9de83f0b-7dd2-4846-a1ce-c8af930778f4\") " pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.287604 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1d080332-c215-44c1-a027-65afbc612f88-signing-key\") pod \"service-ca-9c57cc56f-gd4gc\" (UID: \"1d080332-c215-44c1-a027-65afbc612f88\") " pod="openshift-service-ca/service-ca-9c57cc56f-gd4gc" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.287638 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5aa6958-e573-4efb-a031-218c62b0bec9-secret-volume\") pod \"collect-profiles-29567805-97bct\" (UID: \"a5aa6958-e573-4efb-a031-218c62b0bec9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.287662 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a49709c7-59e0-440e-89c2-177c42cd28e8-apiservice-cert\") pod \"packageserver-d55dfcdfc-sp69c\" (UID: \"a49709c7-59e0-440e-89c2-177c42cd28e8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.287681 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07720715-40fb-4f7d-8f6b-381dc7cf9ea2-config\") pod \"etcd-operator-b45778765-4nfs6\" (UID: \"07720715-40fb-4f7d-8f6b-381dc7cf9ea2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.289259 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/68d61e65-8275-4862-9dae-a75029889b2a-proxy-tls\") pod \"machine-config-controller-84d6567774-57cxh\" (UID: \"68d61e65-8275-4862-9dae-a75029889b2a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-57cxh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.289865 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b0b465ff-f8f3-4a99-9235-1ddc4ce093e4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-v296h\" (UID: \"b0b465ff-f8f3-4a99-9235-1ddc4ce093e4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v296h" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.290590 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e5224539-6d29-4bc3-9656-4665eb287e28-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-d7xpn\" (UID: \"e5224539-6d29-4bc3-9656-4665eb287e28\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-d7xpn" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.290928 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eddac2a3-8b07-410d-9e9f-f7ce0c05abf5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-725jh\" (UID: \"eddac2a3-8b07-410d-9e9f-f7ce0c05abf5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-725jh" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.291572 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5aa6958-e573-4efb-a031-218c62b0bec9-secret-volume\") pod \"collect-profiles-29567805-97bct\" (UID: \"a5aa6958-e573-4efb-a031-218c62b0bec9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.300019 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.319667 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.339961 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.349532 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c0dc3b9-8e11-4ff0-8f26-549ce47215f0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-blwmr\" (UID: \"0c0dc3b9-8e11-4ff0-8f26-549ce47215f0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-blwmr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.360155 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.380694 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.401388 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.420603 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.428066 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c0dc3b9-8e11-4ff0-8f26-549ce47215f0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-blwmr\" (UID: \"0c0dc3b9-8e11-4ff0-8f26-549ce47215f0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-blwmr" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.441396 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.450520 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/bc2b8485-8cc1-4029-8318-397d4278e455-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-mbgvc\" (UID: \"bc2b8485-8cc1-4029-8318-397d4278e455\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mbgvc" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.461220 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.473765 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b0b465ff-f8f3-4a99-9235-1ddc4ce093e4-srv-cert\") pod \"olm-operator-6b444d44fb-v296h\" (UID: \"b0b465ff-f8f3-4a99-9235-1ddc4ce093e4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v296h" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.480303 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.501854 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.520706 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.531622 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1d080332-c215-44c1-a027-65afbc612f88-signing-key\") pod \"service-ca-9c57cc56f-gd4gc\" (UID: \"1d080332-c215-44c1-a027-65afbc612f88\") " pod="openshift-service-ca/service-ca-9c57cc56f-gd4gc" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.540538 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.546930 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1d080332-c215-44c1-a027-65afbc612f88-signing-cabundle\") pod \"service-ca-9c57cc56f-gd4gc\" (UID: \"1d080332-c215-44c1-a027-65afbc612f88\") " pod="openshift-service-ca/service-ca-9c57cc56f-gd4gc" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.560653 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.599827 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.619957 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.639928 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.660708 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.669900 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dfbaac71-f99c-4373-a469-f2e5dd0ee632-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wrzs4\" (UID: \"dfbaac71-f99c-4373-a469-f2e5dd0ee632\") " pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.686111 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.687918 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dfbaac71-f99c-4373-a469-f2e5dd0ee632-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wrzs4\" (UID: \"dfbaac71-f99c-4373-a469-f2e5dd0ee632\") " pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.701391 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.720076 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.728468 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07720715-40fb-4f7d-8f6b-381dc7cf9ea2-config\") pod \"etcd-operator-b45778765-4nfs6\" (UID: \"07720715-40fb-4f7d-8f6b-381dc7cf9ea2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.739897 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.759725 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.769015 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07720715-40fb-4f7d-8f6b-381dc7cf9ea2-serving-cert\") pod \"etcd-operator-b45778765-4nfs6\" (UID: \"07720715-40fb-4f7d-8f6b-381dc7cf9ea2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.779667 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.788524 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/07720715-40fb-4f7d-8f6b-381dc7cf9ea2-etcd-ca\") pod \"etcd-operator-b45778765-4nfs6\" (UID: \"07720715-40fb-4f7d-8f6b-381dc7cf9ea2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.800014 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.807177 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/07720715-40fb-4f7d-8f6b-381dc7cf9ea2-etcd-service-ca\") pod \"etcd-operator-b45778765-4nfs6\" (UID: \"07720715-40fb-4f7d-8f6b-381dc7cf9ea2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.819596 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.835560 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/07720715-40fb-4f7d-8f6b-381dc7cf9ea2-etcd-client\") pod \"etcd-operator-b45778765-4nfs6\" (UID: \"07720715-40fb-4f7d-8f6b-381dc7cf9ea2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.840320 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.860308 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.880789 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.900101 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.910018 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/811f064c-ebf4-48ad-87a0-83205eb1eca5-metrics-tls\") pod \"dns-operator-744455d44c-vjv4q\" (UID: \"811f064c-ebf4-48ad-87a0-83205eb1eca5\") " pod="openshift-dns-operator/dns-operator-744455d44c-vjv4q" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.919787 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.939700 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.960174 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.980942 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 21 04:51:01 crc kubenswrapper[4775]: I0321 04:51:01.992204 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/918078e1-af38-475a-86d5-8179cafa18db-serving-cert\") pod \"service-ca-operator-777779d784-wqjnj\" (UID: \"918078e1-af38-475a-86d5-8179cafa18db\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wqjnj" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.000158 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.008890 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/918078e1-af38-475a-86d5-8179cafa18db-config\") pod \"service-ca-operator-777779d784-wqjnj\" (UID: \"918078e1-af38-475a-86d5-8179cafa18db\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wqjnj" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.020395 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.040753 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.063882 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.080186 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.089300 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/02b06614-31da-49d9-bc97-cf61b065d42f-images\") pod \"machine-config-operator-74547568cd-8kpqj\" (UID: \"02b06614-31da-49d9-bc97-cf61b065d42f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.100427 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.108738 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/02b06614-31da-49d9-bc97-cf61b065d42f-proxy-tls\") pod \"machine-config-operator-74547568cd-8kpqj\" (UID: \"02b06614-31da-49d9-bc97-cf61b065d42f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.118317 4775 request.go:700] Waited for 1.001811711s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-operator-dockercfg-98p87&limit=500&resourceVersion=0 Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.120237 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.140947 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.150346 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a49709c7-59e0-440e-89c2-177c42cd28e8-webhook-cert\") pod \"packageserver-d55dfcdfc-sp69c\" (UID: \"a49709c7-59e0-440e-89c2-177c42cd28e8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.150689 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a49709c7-59e0-440e-89c2-177c42cd28e8-apiservice-cert\") pod \"packageserver-d55dfcdfc-sp69c\" (UID: \"a49709c7-59e0-440e-89c2-177c42cd28e8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.160580 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.179993 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.187077 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5aa6958-e573-4efb-a031-218c62b0bec9-config-volume\") pod \"collect-profiles-29567805-97bct\" (UID: \"a5aa6958-e573-4efb-a031-218c62b0bec9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.200055 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.220453 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.231320 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a79bbac6-f40a-4c92-8854-7ab5e72573cc-default-certificate\") pod \"router-default-5444994796-2z46g\" (UID: \"a79bbac6-f40a-4c92-8854-7ab5e72573cc\") " pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.240841 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.250370 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a79bbac6-f40a-4c92-8854-7ab5e72573cc-stats-auth\") pod \"router-default-5444994796-2z46g\" (UID: \"a79bbac6-f40a-4c92-8854-7ab5e72573cc\") " pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.260337 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.270197 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a79bbac6-f40a-4c92-8854-7ab5e72573cc-metrics-certs\") pod \"router-default-5444994796-2z46g\" (UID: \"a79bbac6-f40a-4c92-8854-7ab5e72573cc\") " pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.280186 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 21 04:51:02 crc kubenswrapper[4775]: E0321 04:51:02.286217 4775 configmap.go:193] Couldn't get configMap openshift-controller-manager-operator/openshift-controller-manager-operator-config: failed to sync configmap cache: timed out waiting for the condition Mar 21 04:51:02 crc kubenswrapper[4775]: E0321 04:51:02.286285 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/edb67e99-0d50-46e9-adbb-c0831dd915d8-config podName:edb67e99-0d50-46e9-adbb-c0831dd915d8 nodeName:}" failed. No retries permitted until 2026-03-21 04:51:02.786269263 +0000 UTC m=+215.762732887 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/edb67e99-0d50-46e9-adbb-c0831dd915d8-config") pod "openshift-controller-manager-operator-756b6f6bc6-xv7nw" (UID: "edb67e99-0d50-46e9-adbb-c0831dd915d8") : failed to sync configmap cache: timed out waiting for the condition Mar 21 04:51:02 crc kubenswrapper[4775]: E0321 04:51:02.286218 4775 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Mar 21 04:51:02 crc kubenswrapper[4775]: E0321 04:51:02.286329 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ae87529c-e52f-45b0-9bbf-2a652e628bc5-metrics-tls podName:ae87529c-e52f-45b0-9bbf-2a652e628bc5 nodeName:}" failed. No retries permitted until 2026-03-21 04:51:02.786319765 +0000 UTC m=+215.762783389 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/ae87529c-e52f-45b0-9bbf-2a652e628bc5-metrics-tls") pod "dns-default-zs5pr" (UID: "ae87529c-e52f-45b0-9bbf-2a652e628bc5") : failed to sync secret cache: timed out waiting for the condition Mar 21 04:51:02 crc kubenswrapper[4775]: E0321 04:51:02.287330 4775 configmap.go:193] Couldn't get configMap openshift-ingress/service-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Mar 21 04:51:02 crc kubenswrapper[4775]: E0321 04:51:02.287406 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a79bbac6-f40a-4c92-8854-7ab5e72573cc-service-ca-bundle podName:a79bbac6-f40a-4c92-8854-7ab5e72573cc nodeName:}" failed. No retries permitted until 2026-03-21 04:51:02.787392725 +0000 UTC m=+215.763856359 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/a79bbac6-f40a-4c92-8854-7ab5e72573cc-service-ca-bundle") pod "router-default-5444994796-2z46g" (UID: "a79bbac6-f40a-4c92-8854-7ab5e72573cc") : failed to sync configmap cache: timed out waiting for the condition Mar 21 04:51:02 crc kubenswrapper[4775]: E0321 04:51:02.287429 4775 secret.go:188] Couldn't get secret openshift-kube-scheduler-operator/kube-scheduler-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 21 04:51:02 crc kubenswrapper[4775]: E0321 04:51:02.287455 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f9395e24-0d4b-4165-bf60-068876927f58-serving-cert podName:f9395e24-0d4b-4165-bf60-068876927f58 nodeName:}" failed. No retries permitted until 2026-03-21 04:51:02.787447307 +0000 UTC m=+215.763911051 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/f9395e24-0d4b-4165-bf60-068876927f58-serving-cert") pod "openshift-kube-scheduler-operator-5fdd9b5758-5ztsd" (UID: "f9395e24-0d4b-4165-bf60-068876927f58") : failed to sync secret cache: timed out waiting for the condition Mar 21 04:51:02 crc kubenswrapper[4775]: E0321 04:51:02.287500 4775 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Mar 21 04:51:02 crc kubenswrapper[4775]: E0321 04:51:02.287546 4775 configmap.go:193] Couldn't get configMap openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-config: failed to sync configmap cache: timed out waiting for the condition Mar 21 04:51:02 crc kubenswrapper[4775]: E0321 04:51:02.287566 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ae87529c-e52f-45b0-9bbf-2a652e628bc5-config-volume podName:ae87529c-e52f-45b0-9bbf-2a652e628bc5 nodeName:}" failed. No retries permitted until 2026-03-21 04:51:02.78755678 +0000 UTC m=+215.764020404 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/ae87529c-e52f-45b0-9bbf-2a652e628bc5-config-volume") pod "dns-default-zs5pr" (UID: "ae87529c-e52f-45b0-9bbf-2a652e628bc5") : failed to sync configmap cache: timed out waiting for the condition Mar 21 04:51:02 crc kubenswrapper[4775]: E0321 04:51:02.287669 4775 secret.go:188] Couldn't get secret openshift-kube-apiserver-operator/kube-apiserver-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 21 04:51:02 crc kubenswrapper[4775]: E0321 04:51:02.287699 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f9395e24-0d4b-4165-bf60-068876927f58-config podName:f9395e24-0d4b-4165-bf60-068876927f58 nodeName:}" failed. No retries permitted until 2026-03-21 04:51:02.787678614 +0000 UTC m=+215.764142238 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/f9395e24-0d4b-4165-bf60-068876927f58-config") pod "openshift-kube-scheduler-operator-5fdd9b5758-5ztsd" (UID: "f9395e24-0d4b-4165-bf60-068876927f58") : failed to sync configmap cache: timed out waiting for the condition Mar 21 04:51:02 crc kubenswrapper[4775]: E0321 04:51:02.287582 4775 secret.go:188] Couldn't get secret openshift-controller-manager-operator/openshift-controller-manager-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 21 04:51:02 crc kubenswrapper[4775]: E0321 04:51:02.287716 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1017f8be-2192-46a4-8717-ead73ca5e81b-serving-cert podName:1017f8be-2192-46a4-8717-ead73ca5e81b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:02.787708925 +0000 UTC m=+215.764172549 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/1017f8be-2192-46a4-8717-ead73ca5e81b-serving-cert") pod "kube-apiserver-operator-766d6c64bb-2zh8p" (UID: "1017f8be-2192-46a4-8717-ead73ca5e81b") : failed to sync secret cache: timed out waiting for the condition Mar 21 04:51:02 crc kubenswrapper[4775]: E0321 04:51:02.287643 4775 configmap.go:193] Couldn't get configMap openshift-kube-apiserver-operator/kube-apiserver-operator-config: failed to sync configmap cache: timed out waiting for the condition Mar 21 04:51:02 crc kubenswrapper[4775]: E0321 04:51:02.287733 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/edb67e99-0d50-46e9-adbb-c0831dd915d8-serving-cert podName:edb67e99-0d50-46e9-adbb-c0831dd915d8 nodeName:}" failed. No retries permitted until 2026-03-21 04:51:02.787724085 +0000 UTC m=+215.764187829 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/edb67e99-0d50-46e9-adbb-c0831dd915d8-serving-cert") pod "openshift-controller-manager-operator-756b6f6bc6-xv7nw" (UID: "edb67e99-0d50-46e9-adbb-c0831dd915d8") : failed to sync secret cache: timed out waiting for the condition Mar 21 04:51:02 crc kubenswrapper[4775]: E0321 04:51:02.287768 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1017f8be-2192-46a4-8717-ead73ca5e81b-config podName:1017f8be-2192-46a4-8717-ead73ca5e81b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:02.787761176 +0000 UTC m=+215.764224800 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/1017f8be-2192-46a4-8717-ead73ca5e81b-config") pod "kube-apiserver-operator-766d6c64bb-2zh8p" (UID: "1017f8be-2192-46a4-8717-ead73ca5e81b") : failed to sync configmap cache: timed out waiting for the condition Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.300297 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.320419 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.340106 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.360806 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.381023 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.400603 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.420541 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.440559 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.459878 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.480954 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.482441 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.482524 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.500234 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.520609 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.540703 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.560761 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.580502 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.606577 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.620034 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.640104 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.661142 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.681430 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.701103 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.720207 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.740854 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.760925 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.782913 4775 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.802009 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.809617 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1017f8be-2192-46a4-8717-ead73ca5e81b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2zh8p\" (UID: \"1017f8be-2192-46a4-8717-ead73ca5e81b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zh8p" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.809691 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a79bbac6-f40a-4c92-8854-7ab5e72573cc-service-ca-bundle\") pod \"router-default-5444994796-2z46g\" (UID: \"a79bbac6-f40a-4c92-8854-7ab5e72573cc\") " pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.809738 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/edb67e99-0d50-46e9-adbb-c0831dd915d8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-xv7nw\" (UID: \"edb67e99-0d50-46e9-adbb-c0831dd915d8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xv7nw" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.809841 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1017f8be-2192-46a4-8717-ead73ca5e81b-config\") pod \"kube-apiserver-operator-766d6c64bb-2zh8p\" (UID: \"1017f8be-2192-46a4-8717-ead73ca5e81b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zh8p" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.810781 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9395e24-0d4b-4165-bf60-068876927f58-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5ztsd\" (UID: \"f9395e24-0d4b-4165-bf60-068876927f58\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5ztsd" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.810831 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae87529c-e52f-45b0-9bbf-2a652e628bc5-config-volume\") pod \"dns-default-zs5pr\" (UID: \"ae87529c-e52f-45b0-9bbf-2a652e628bc5\") " pod="openshift-dns/dns-default-zs5pr" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.810851 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9395e24-0d4b-4165-bf60-068876927f58-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5ztsd\" (UID: \"f9395e24-0d4b-4165-bf60-068876927f58\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5ztsd" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.810917 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edb67e99-0d50-46e9-adbb-c0831dd915d8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-xv7nw\" (UID: \"edb67e99-0d50-46e9-adbb-c0831dd915d8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xv7nw" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.811281 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a79bbac6-f40a-4c92-8854-7ab5e72573cc-service-ca-bundle\") pod \"router-default-5444994796-2z46g\" (UID: \"a79bbac6-f40a-4c92-8854-7ab5e72573cc\") " pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.811381 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae87529c-e52f-45b0-9bbf-2a652e628bc5-metrics-tls\") pod \"dns-default-zs5pr\" (UID: \"ae87529c-e52f-45b0-9bbf-2a652e628bc5\") " pod="openshift-dns/dns-default-zs5pr" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.812021 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1017f8be-2192-46a4-8717-ead73ca5e81b-config\") pod \"kube-apiserver-operator-766d6c64bb-2zh8p\" (UID: \"1017f8be-2192-46a4-8717-ead73ca5e81b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zh8p" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.812344 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae87529c-e52f-45b0-9bbf-2a652e628bc5-config-volume\") pod \"dns-default-zs5pr\" (UID: \"ae87529c-e52f-45b0-9bbf-2a652e628bc5\") " pod="openshift-dns/dns-default-zs5pr" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.812349 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9395e24-0d4b-4165-bf60-068876927f58-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5ztsd\" (UID: \"f9395e24-0d4b-4165-bf60-068876927f58\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5ztsd" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.812617 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edb67e99-0d50-46e9-adbb-c0831dd915d8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-xv7nw\" (UID: \"edb67e99-0d50-46e9-adbb-c0831dd915d8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xv7nw" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.815088 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/edb67e99-0d50-46e9-adbb-c0831dd915d8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-xv7nw\" (UID: \"edb67e99-0d50-46e9-adbb-c0831dd915d8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xv7nw" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.815276 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9395e24-0d4b-4165-bf60-068876927f58-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5ztsd\" (UID: \"f9395e24-0d4b-4165-bf60-068876927f58\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5ztsd" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.815456 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae87529c-e52f-45b0-9bbf-2a652e628bc5-metrics-tls\") pod \"dns-default-zs5pr\" (UID: \"ae87529c-e52f-45b0-9bbf-2a652e628bc5\") " pod="openshift-dns/dns-default-zs5pr" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.817771 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1017f8be-2192-46a4-8717-ead73ca5e81b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2zh8p\" (UID: \"1017f8be-2192-46a4-8717-ead73ca5e81b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zh8p" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.857720 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k97vh\" (UniqueName: \"kubernetes.io/projected/fe3df1e1-4c22-48df-aaea-469c864f0310-kube-api-access-k97vh\") pod \"machine-api-operator-5694c8668f-5rnj6\" (UID: \"fe3df1e1-4c22-48df-aaea-469c864f0310\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5rnj6" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.858974 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.883775 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhsgq\" (UniqueName: \"kubernetes.io/projected/c3d58eba-4ddf-463c-baa1-1943fb60c732-kube-api-access-bhsgq\") pod \"downloads-7954f5f757-nstnr\" (UID: \"c3d58eba-4ddf-463c-baa1-1943fb60c732\") " pod="openshift-console/downloads-7954f5f757-nstnr" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.896745 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zb5h\" (UniqueName: \"kubernetes.io/projected/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-kube-api-access-2zb5h\") pod \"controller-manager-879f6c89f-r6c6z\" (UID: \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.914010 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntrqw\" (UniqueName: \"kubernetes.io/projected/768c9343-1391-4221-b008-5dee2921953f-kube-api-access-ntrqw\") pod \"authentication-operator-69f744f599-dzxd5\" (UID: \"768c9343-1391-4221-b008-5dee2921953f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dzxd5" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.934415 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pflfp\" (UniqueName: \"kubernetes.io/projected/6343dd4f-5c5b-4c94-a4d6-f603698ba6ac-kube-api-access-pflfp\") pod \"machine-approver-56656f9798-468g4\" (UID: \"6343dd4f-5c5b-4c94-a4d6-f603698ba6ac\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-468g4" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.942668 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.966777 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhtwp\" (UniqueName: \"kubernetes.io/projected/892d15b6-460e-4892-a836-0cc284c8a326-kube-api-access-mhtwp\") pod \"openshift-config-operator-7777fb866f-mthhs\" (UID: \"892d15b6-460e-4892-a836-0cc284c8a326\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.975480 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fw6r\" (UniqueName: \"kubernetes.io/projected/fddde3da-8512-4e62-9c38-b59f98e117e0-kube-api-access-4fw6r\") pod \"oauth-openshift-558db77b4-jd86b\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.989612 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-5rnj6" Mar 21 04:51:02 crc kubenswrapper[4775]: I0321 04:51:02.994447 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/10a9b5d9-308c-4971-a073-9c88de98d8ea-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9v866\" (UID: \"10a9b5d9-308c-4971-a073-9c88de98d8ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9v866" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.022678 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-nstnr" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.032472 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-468g4" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.053444 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvbpw\" (UniqueName: \"kubernetes.io/projected/edb4245a-7971-4c40-81b6-27d56b319a2f-kube-api-access-dvbpw\") pod \"cluster-samples-operator-665b6dd947-k5dks\" (UID: \"edb4245a-7971-4c40-81b6-27d56b319a2f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k5dks" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.062799 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqkw4\" (UniqueName: \"kubernetes.io/projected/8bb7828f-6d99-4539-8312-c8e96bfbc608-kube-api-access-lqkw4\") pod \"apiserver-7bbb656c7d-kxtlf\" (UID: \"8bb7828f-6d99-4539-8312-c8e96bfbc608\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.100901 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nbfq\" (UniqueName: \"kubernetes.io/projected/2d2512b8-9fb6-4b9b-9ac2-268e0b8bd04a-kube-api-access-7nbfq\") pod \"openshift-apiserver-operator-796bbdcf4f-5pdcr\" (UID: \"2d2512b8-9fb6-4b9b-9ac2-268e0b8bd04a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5pdcr" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.101503 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m6fm\" (UniqueName: \"kubernetes.io/projected/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-kube-api-access-8m6fm\") pod \"console-f9d7485db-wst2s\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.111986 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk47r\" (UniqueName: \"kubernetes.io/projected/12b1cc27-6a60-43d2-9d5e-eb7a54c1e899-kube-api-access-rk47r\") pod \"console-operator-58897d9998-fc4z8\" (UID: \"12b1cc27-6a60-43d2-9d5e-eb7a54c1e899\") " pod="openshift-console-operator/console-operator-58897d9998-fc4z8" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.118623 4775 request.go:700] Waited for 1.938755618s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/serviceaccounts/cluster-image-registry-operator/token Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.126681 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7h98\" (UniqueName: \"kubernetes.io/projected/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-kube-api-access-r7h98\") pod \"route-controller-manager-6576b87f9c-55lcp\" (UID: \"ed9baf4c-70b2-450f-9b21-76dfafbc44d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.138579 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-td4ls\" (UniqueName: \"kubernetes.io/projected/10a9b5d9-308c-4971-a073-9c88de98d8ea-kube-api-access-td4ls\") pod \"cluster-image-registry-operator-dc59b4c8b-9v866\" (UID: \"10a9b5d9-308c-4971-a073-9c88de98d8ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9v866" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.153731 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r6c6z"] Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.156779 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.159546 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.160869 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75ck2\" (UniqueName: \"kubernetes.io/projected/6e22b5c7-4191-4f21-82ba-3014ccc4e978-kube-api-access-75ck2\") pod \"apiserver-76f77b778f-rz6g5\" (UID: \"6e22b5c7-4191-4f21-82ba-3014ccc4e978\") " pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.166930 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.181364 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.196393 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-dzxd5" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.203485 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.210106 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.223337 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.238182 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-fc4z8" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.239023 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5rnj6"] Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.240196 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.249445 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-468g4" event={"ID":"6343dd4f-5c5b-4c94-a4d6-f603698ba6ac","Type":"ContainerStarted","Data":"17aca1ec1c735ebefc774f0f3540a58335b82734921e82f6cf0312e200950325"} Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.250172 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" event={"ID":"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb","Type":"ContainerStarted","Data":"3aa7cdf6ad840aefd2de6cc5405b71b516de12c087774e2fad897251db2fc6ea"} Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.258172 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.259388 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.259849 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.277804 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5pdcr" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.284304 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.298189 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.319406 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m44cw\" (UniqueName: \"kubernetes.io/projected/dfbaac71-f99c-4373-a469-f2e5dd0ee632-kube-api-access-m44cw\") pod \"marketplace-operator-79b997595-wrzs4\" (UID: \"dfbaac71-f99c-4373-a469-f2e5dd0ee632\") " pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.320330 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9v866" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.327402 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k5dks" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.338717 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-td8g7\" (UniqueName: \"kubernetes.io/projected/ae87529c-e52f-45b0-9bbf-2a652e628bc5-kube-api-access-td8g7\") pod \"dns-default-zs5pr\" (UID: \"ae87529c-e52f-45b0-9bbf-2a652e628bc5\") " pod="openshift-dns/dns-default-zs5pr" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.362457 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r82zp\" (UniqueName: \"kubernetes.io/projected/918078e1-af38-475a-86d5-8179cafa18db-kube-api-access-r82zp\") pod \"service-ca-operator-777779d784-wqjnj\" (UID: \"918078e1-af38-475a-86d5-8179cafa18db\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wqjnj" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.388276 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpvcx\" (UniqueName: \"kubernetes.io/projected/811f064c-ebf4-48ad-87a0-83205eb1eca5-kube-api-access-kpvcx\") pod \"dns-operator-744455d44c-vjv4q\" (UID: \"811f064c-ebf4-48ad-87a0-83205eb1eca5\") " pod="openshift-dns-operator/dns-operator-744455d44c-vjv4q" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.397929 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f9395e24-0d4b-4165-bf60-068876927f58-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5ztsd\" (UID: \"f9395e24-0d4b-4165-bf60-068876927f58\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5ztsd" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.418939 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clmnm\" (UniqueName: \"kubernetes.io/projected/9de83f0b-7dd2-4846-a1ce-c8af930778f4-kube-api-access-clmnm\") pod \"csi-hostpathplugin-hvfsp\" (UID: \"9de83f0b-7dd2-4846-a1ce-c8af930778f4\") " pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.433287 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.439269 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9mb9\" (UniqueName: \"kubernetes.io/projected/edb67e99-0d50-46e9-adbb-c0831dd915d8-kube-api-access-d9mb9\") pod \"openshift-controller-manager-operator-756b6f6bc6-xv7nw\" (UID: \"edb67e99-0d50-46e9-adbb-c0831dd915d8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xv7nw" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.454895 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vjv4q" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.456322 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wqjnj" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.457406 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nw58z\" (UniqueName: \"kubernetes.io/projected/07720715-40fb-4f7d-8f6b-381dc7cf9ea2-kube-api-access-nw58z\") pod \"etcd-operator-b45778765-4nfs6\" (UID: \"07720715-40fb-4f7d-8f6b-381dc7cf9ea2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.501528 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plx8q\" (UniqueName: \"kubernetes.io/projected/b0b465ff-f8f3-4a99-9235-1ddc4ce093e4-kube-api-access-plx8q\") pod \"olm-operator-6b444d44fb-v296h\" (UID: \"b0b465ff-f8f3-4a99-9235-1ddc4ce093e4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v296h" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.521979 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zfpw\" (UniqueName: \"kubernetes.io/projected/e5224539-6d29-4bc3-9656-4665eb287e28-kube-api-access-7zfpw\") pod \"control-plane-machine-set-operator-78cbb6b69f-d7xpn\" (UID: \"e5224539-6d29-4bc3-9656-4665eb287e28\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-d7xpn" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.528244 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xv7nw" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.540402 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5ztsd" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.545200 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-nstnr"] Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.549891 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5fwg\" (UniqueName: \"kubernetes.io/projected/02b06614-31da-49d9-bc97-cf61b065d42f-kube-api-access-s5fwg\") pod \"machine-config-operator-74547568cd-8kpqj\" (UID: \"02b06614-31da-49d9-bc97-cf61b065d42f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.551590 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gr9p\" (UniqueName: \"kubernetes.io/projected/2dbc150e-3a25-4b44-b01c-effe99de5152-kube-api-access-5gr9p\") pod \"migrator-59844c95c7-7sfkh\" (UID: \"2dbc150e-3a25-4b44-b01c-effe99de5152\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7sfkh" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.556462 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-zs5pr" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.566412 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.567949 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h4n4\" (UniqueName: \"kubernetes.io/projected/68d61e65-8275-4862-9dae-a75029889b2a-kube-api-access-8h4n4\") pod \"machine-config-controller-84d6567774-57cxh\" (UID: \"68d61e65-8275-4862-9dae-a75029889b2a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-57cxh" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.589195 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm9fk\" (UniqueName: \"kubernetes.io/projected/a5aa6958-e573-4efb-a031-218c62b0bec9-kube-api-access-wm9fk\") pod \"collect-profiles-29567805-97bct\" (UID: \"a5aa6958-e573-4efb-a031-218c62b0bec9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.602957 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqpfp\" (UniqueName: \"kubernetes.io/projected/bc2b8485-8cc1-4029-8318-397d4278e455-kube-api-access-tqpfp\") pod \"package-server-manager-789f6589d5-mbgvc\" (UID: \"bc2b8485-8cc1-4029-8318-397d4278e455\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mbgvc" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.619964 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c52b4\" (UniqueName: \"kubernetes.io/projected/0c0dc3b9-8e11-4ff0-8f26-549ce47215f0-kube-api-access-c52b4\") pod \"kube-storage-version-migrator-operator-b67b599dd-blwmr\" (UID: \"0c0dc3b9-8e11-4ff0-8f26-549ce47215f0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-blwmr" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.640365 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-57cxh" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.643138 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vd8k\" (UniqueName: \"kubernetes.io/projected/1d080332-c215-44c1-a027-65afbc612f88-kube-api-access-6vd8k\") pod \"service-ca-9c57cc56f-gd4gc\" (UID: \"1d080332-c215-44c1-a027-65afbc612f88\") " pod="openshift-service-ca/service-ca-9c57cc56f-gd4gc" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.664137 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eddac2a3-8b07-410d-9e9f-f7ce0c05abf5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-725jh\" (UID: \"eddac2a3-8b07-410d-9e9f-f7ce0c05abf5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-725jh" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.664296 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7sfkh" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.678938 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-d7xpn" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.679379 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-725jh" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.687818 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1017f8be-2192-46a4-8717-ead73ca5e81b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2zh8p\" (UID: \"1017f8be-2192-46a4-8717-ead73ca5e81b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zh8p" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.697533 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-blwmr" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.705376 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mbgvc" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.712354 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v296h" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.712557 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhqjf\" (UniqueName: \"kubernetes.io/projected/a79bbac6-f40a-4c92-8854-7ab5e72573cc-kube-api-access-nhqjf\") pod \"router-default-5444994796-2z46g\" (UID: \"a79bbac6-f40a-4c92-8854-7ab5e72573cc\") " pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.714619 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm89b\" (UniqueName: \"kubernetes.io/projected/cd25c8a4-8047-4602-a95b-3308af65bd38-kube-api-access-gm89b\") pod \"auto-csr-approver-29567810-wb89g\" (UID: \"cd25c8a4-8047-4602-a95b-3308af65bd38\") " pod="openshift-infra/auto-csr-approver-29567810-wb89g" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.724794 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-gd4gc" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.737820 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.747565 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9blnw\" (UniqueName: \"kubernetes.io/projected/a49709c7-59e0-440e-89c2-177c42cd28e8-kube-api-access-9blnw\") pod \"packageserver-d55dfcdfc-sp69c\" (UID: \"a49709c7-59e0-440e-89c2-177c42cd28e8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.761148 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf"] Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.777746 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567810-wb89g" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.790278 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dzxd5"] Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.809395 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.809616 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.809906 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.813524 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.819437 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zh8p" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.835708 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.835779 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-registry-tls\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.835815 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d55e67e9-9980-4878-9875-a07207894f6f-profile-collector-cert\") pod \"catalog-operator-68c6474976-zkl7p\" (UID: \"d55e67e9-9980-4878-9875-a07207894f6f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zkl7p" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.835834 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-bound-sa-token\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.835861 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqdww\" (UniqueName: \"kubernetes.io/projected/f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8-kube-api-access-bqdww\") pod \"ingress-operator-5b745b69d9-vsh8z\" (UID: \"f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.835896 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.835926 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2j66\" (UniqueName: \"kubernetes.io/projected/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-kube-api-access-z2j66\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.835955 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8-trusted-ca\") pod \"ingress-operator-5b745b69d9-vsh8z\" (UID: \"f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.835987 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.836006 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b0a24bc9-b395-4732-ba57-b096ee9ffdb1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2lnkp\" (UID: \"b0a24bc9-b395-4732-ba57-b096ee9ffdb1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2lnkp" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.836034 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8-metrics-tls\") pod \"ingress-operator-5b745b69d9-vsh8z\" (UID: \"f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.836082 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d55e67e9-9980-4878-9875-a07207894f6f-srv-cert\") pod \"catalog-operator-68c6474976-zkl7p\" (UID: \"d55e67e9-9980-4878-9875-a07207894f6f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zkl7p" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.836135 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv7kh\" (UniqueName: \"kubernetes.io/projected/d55e67e9-9980-4878-9875-a07207894f6f-kube-api-access-hv7kh\") pod \"catalog-operator-68c6474976-zkl7p\" (UID: \"d55e67e9-9980-4878-9875-a07207894f6f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zkl7p" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.836182 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhcjp\" (UniqueName: \"kubernetes.io/projected/b0a24bc9-b395-4732-ba57-b096ee9ffdb1-kube-api-access-hhcjp\") pod \"multus-admission-controller-857f4d67dd-2lnkp\" (UID: \"b0a24bc9-b395-4732-ba57-b096ee9ffdb1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2lnkp" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.836215 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-registry-certificates\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.836232 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-trusted-ca\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.836272 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vsh8z\" (UID: \"f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z" Mar 21 04:51:03 crc kubenswrapper[4775]: E0321 04:51:03.839652 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:04.339637433 +0000 UTC m=+217.316101057 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:03 crc kubenswrapper[4775]: W0321 04:51:03.896796 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda79bbac6_f40a_4c92_8854_7ab5e72573cc.slice/crio-20eb94421e0eea64be64321e6c00b000dee79fd4a4662f6a369022dd15653f7e WatchSource:0}: Error finding container 20eb94421e0eea64be64321e6c00b000dee79fd4a4662f6a369022dd15653f7e: Status 404 returned error can't find the container with id 20eb94421e0eea64be64321e6c00b000dee79fd4a4662f6a369022dd15653f7e Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.937734 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.938285 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.938347 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-registry-tls\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.938366 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d55e67e9-9980-4878-9875-a07207894f6f-profile-collector-cert\") pod \"catalog-operator-68c6474976-zkl7p\" (UID: \"d55e67e9-9980-4878-9875-a07207894f6f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zkl7p" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.938418 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-bound-sa-token\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.938470 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqdww\" (UniqueName: \"kubernetes.io/projected/f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8-kube-api-access-bqdww\") pod \"ingress-operator-5b745b69d9-vsh8z\" (UID: \"f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.938516 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/ecae58ef-adaa-42ea-9246-49845334a819-node-bootstrap-token\") pod \"machine-config-server-dnv7h\" (UID: \"ecae58ef-adaa-42ea-9246-49845334a819\") " pod="openshift-machine-config-operator/machine-config-server-dnv7h" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.938552 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.938594 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/35c6eec5-7117-48de-a4c5-640517da6a76-cert\") pod \"ingress-canary-6jnn5\" (UID: \"35c6eec5-7117-48de-a4c5-640517da6a76\") " pod="openshift-ingress-canary/ingress-canary-6jnn5" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.938681 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2j66\" (UniqueName: \"kubernetes.io/projected/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-kube-api-access-z2j66\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.938731 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8-trusted-ca\") pod \"ingress-operator-5b745b69d9-vsh8z\" (UID: \"f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.938772 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b0a24bc9-b395-4732-ba57-b096ee9ffdb1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2lnkp\" (UID: \"b0a24bc9-b395-4732-ba57-b096ee9ffdb1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2lnkp" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.938841 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8-metrics-tls\") pod \"ingress-operator-5b745b69d9-vsh8z\" (UID: \"f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.938927 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d55e67e9-9980-4878-9875-a07207894f6f-srv-cert\") pod \"catalog-operator-68c6474976-zkl7p\" (UID: \"d55e67e9-9980-4878-9875-a07207894f6f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zkl7p" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.939031 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv7kh\" (UniqueName: \"kubernetes.io/projected/d55e67e9-9980-4878-9875-a07207894f6f-kube-api-access-hv7kh\") pod \"catalog-operator-68c6474976-zkl7p\" (UID: \"d55e67e9-9980-4878-9875-a07207894f6f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zkl7p" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.939085 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhcjp\" (UniqueName: \"kubernetes.io/projected/b0a24bc9-b395-4732-ba57-b096ee9ffdb1-kube-api-access-hhcjp\") pod \"multus-admission-controller-857f4d67dd-2lnkp\" (UID: \"b0a24bc9-b395-4732-ba57-b096ee9ffdb1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2lnkp" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.939205 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-registry-certificates\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.939241 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-trusted-ca\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.939265 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbgjp\" (UniqueName: \"kubernetes.io/projected/ecae58ef-adaa-42ea-9246-49845334a819-kube-api-access-xbgjp\") pod \"machine-config-server-dnv7h\" (UID: \"ecae58ef-adaa-42ea-9246-49845334a819\") " pod="openshift-machine-config-operator/machine-config-server-dnv7h" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.939302 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vsh8z\" (UID: \"f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.939331 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/ecae58ef-adaa-42ea-9246-49845334a819-certs\") pod \"machine-config-server-dnv7h\" (UID: \"ecae58ef-adaa-42ea-9246-49845334a819\") " pod="openshift-machine-config-operator/machine-config-server-dnv7h" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.939369 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bpwh\" (UniqueName: \"kubernetes.io/projected/35c6eec5-7117-48de-a4c5-640517da6a76-kube-api-access-5bpwh\") pod \"ingress-canary-6jnn5\" (UID: \"35c6eec5-7117-48de-a4c5-640517da6a76\") " pod="openshift-ingress-canary/ingress-canary-6jnn5" Mar 21 04:51:03 crc kubenswrapper[4775]: E0321 04:51:03.945076 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:04.441353055 +0000 UTC m=+217.417816689 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.954074 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8-trusted-ca\") pod \"ingress-operator-5b745b69d9-vsh8z\" (UID: \"f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.954343 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.955175 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8-metrics-tls\") pod \"ingress-operator-5b745b69d9-vsh8z\" (UID: \"f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.955544 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b0a24bc9-b395-4732-ba57-b096ee9ffdb1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2lnkp\" (UID: \"b0a24bc9-b395-4732-ba57-b096ee9ffdb1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2lnkp" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.965800 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-trusted-ca\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.970405 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqdww\" (UniqueName: \"kubernetes.io/projected/f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8-kube-api-access-bqdww\") pod \"ingress-operator-5b745b69d9-vsh8z\" (UID: \"f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.972636 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d55e67e9-9980-4878-9875-a07207894f6f-srv-cert\") pod \"catalog-operator-68c6474976-zkl7p\" (UID: \"d55e67e9-9980-4878-9875-a07207894f6f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zkl7p" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.975419 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-registry-tls\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.976193 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-bound-sa-token\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.976840 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-registry-certificates\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:03 crc kubenswrapper[4775]: I0321 04:51:03.977889 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d55e67e9-9980-4878-9875-a07207894f6f-profile-collector-cert\") pod \"catalog-operator-68c6474976-zkl7p\" (UID: \"d55e67e9-9980-4878-9875-a07207894f6f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zkl7p" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.008399 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.040413 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2j66\" (UniqueName: \"kubernetes.io/projected/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-kube-api-access-z2j66\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.042075 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/ecae58ef-adaa-42ea-9246-49845334a819-node-bootstrap-token\") pod \"machine-config-server-dnv7h\" (UID: \"ecae58ef-adaa-42ea-9246-49845334a819\") " pod="openshift-machine-config-operator/machine-config-server-dnv7h" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.042110 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/35c6eec5-7117-48de-a4c5-640517da6a76-cert\") pod \"ingress-canary-6jnn5\" (UID: \"35c6eec5-7117-48de-a4c5-640517da6a76\") " pod="openshift-ingress-canary/ingress-canary-6jnn5" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.042166 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.042224 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbgjp\" (UniqueName: \"kubernetes.io/projected/ecae58ef-adaa-42ea-9246-49845334a819-kube-api-access-xbgjp\") pod \"machine-config-server-dnv7h\" (UID: \"ecae58ef-adaa-42ea-9246-49845334a819\") " pod="openshift-machine-config-operator/machine-config-server-dnv7h" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.042245 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/ecae58ef-adaa-42ea-9246-49845334a819-certs\") pod \"machine-config-server-dnv7h\" (UID: \"ecae58ef-adaa-42ea-9246-49845334a819\") " pod="openshift-machine-config-operator/machine-config-server-dnv7h" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.042262 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bpwh\" (UniqueName: \"kubernetes.io/projected/35c6eec5-7117-48de-a4c5-640517da6a76-kube-api-access-5bpwh\") pod \"ingress-canary-6jnn5\" (UID: \"35c6eec5-7117-48de-a4c5-640517da6a76\") " pod="openshift-ingress-canary/ingress-canary-6jnn5" Mar 21 04:51:04 crc kubenswrapper[4775]: E0321 04:51:04.042625 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:04.542614213 +0000 UTC m=+217.519077837 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.049090 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/35c6eec5-7117-48de-a4c5-640517da6a76-cert\") pod \"ingress-canary-6jnn5\" (UID: \"35c6eec5-7117-48de-a4c5-640517da6a76\") " pod="openshift-ingress-canary/ingress-canary-6jnn5" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.049506 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/ecae58ef-adaa-42ea-9246-49845334a819-node-bootstrap-token\") pod \"machine-config-server-dnv7h\" (UID: \"ecae58ef-adaa-42ea-9246-49845334a819\") " pod="openshift-machine-config-operator/machine-config-server-dnv7h" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.051981 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv7kh\" (UniqueName: \"kubernetes.io/projected/d55e67e9-9980-4878-9875-a07207894f6f-kube-api-access-hv7kh\") pod \"catalog-operator-68c6474976-zkl7p\" (UID: \"d55e67e9-9980-4878-9875-a07207894f6f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zkl7p" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.054515 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/ecae58ef-adaa-42ea-9246-49845334a819-certs\") pod \"machine-config-server-dnv7h\" (UID: \"ecae58ef-adaa-42ea-9246-49845334a819\") " pod="openshift-machine-config-operator/machine-config-server-dnv7h" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.059812 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhcjp\" (UniqueName: \"kubernetes.io/projected/b0a24bc9-b395-4732-ba57-b096ee9ffdb1-kube-api-access-hhcjp\") pod \"multus-admission-controller-857f4d67dd-2lnkp\" (UID: \"b0a24bc9-b395-4732-ba57-b096ee9ffdb1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2lnkp" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.089707 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vsh8z\" (UID: \"f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.119823 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbgjp\" (UniqueName: \"kubernetes.io/projected/ecae58ef-adaa-42ea-9246-49845334a819-kube-api-access-xbgjp\") pod \"machine-config-server-dnv7h\" (UID: \"ecae58ef-adaa-42ea-9246-49845334a819\") " pod="openshift-machine-config-operator/machine-config-server-dnv7h" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.145617 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:04 crc kubenswrapper[4775]: E0321 04:51:04.146138 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:04.646100044 +0000 UTC m=+217.622563668 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.147865 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.159546 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bpwh\" (UniqueName: \"kubernetes.io/projected/35c6eec5-7117-48de-a4c5-640517da6a76-kube-api-access-5bpwh\") pod \"ingress-canary-6jnn5\" (UID: \"35c6eec5-7117-48de-a4c5-640517da6a76\") " pod="openshift-ingress-canary/ingress-canary-6jnn5" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.191705 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-6jnn5" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.201682 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-dnv7h" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.248878 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:04 crc kubenswrapper[4775]: E0321 04:51:04.249171 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:04.749158473 +0000 UTC m=+217.725622097 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.256823 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zkl7p" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.263884 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-2lnkp" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.290913 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-fc4z8"] Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.292666 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-nstnr" event={"ID":"c3d58eba-4ddf-463c-baa1-1943fb60c732","Type":"ContainerStarted","Data":"591e37f237061eb3eb2e78ab869ffea4d983607a257bac29e4fb969c2fcd9940"} Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.292707 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-nstnr" event={"ID":"c3d58eba-4ddf-463c-baa1-1943fb60c732","Type":"ContainerStarted","Data":"5523732cc967700d7fbbdac853665f17cf7feebe5b7b0a8fb0f722ef5753d557"} Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.292992 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-nstnr" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.294082 4775 patch_prober.go:28] interesting pod/downloads-7954f5f757-nstnr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.294198 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nstnr" podUID="c3d58eba-4ddf-463c-baa1-1943fb60c732" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.297680 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5rnj6" event={"ID":"fe3df1e1-4c22-48df-aaea-469c864f0310","Type":"ContainerStarted","Data":"9544447e99441f87401ed20c2711c92e4aea86d65db8c4871e6c026e52807c41"} Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.297963 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5rnj6" event={"ID":"fe3df1e1-4c22-48df-aaea-469c864f0310","Type":"ContainerStarted","Data":"d9f6011b3900db1117e47de329dbbfd8cf7a8f07048ae0cad95b86eb83444093"} Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.297978 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5rnj6" event={"ID":"fe3df1e1-4c22-48df-aaea-469c864f0310","Type":"ContainerStarted","Data":"b7eff7cc15d1d879cb238a4a50f9e6b0725bb469b6c1561f9ac5294d7c2bc0c6"} Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.303231 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-468g4" event={"ID":"6343dd4f-5c5b-4c94-a4d6-f603698ba6ac","Type":"ContainerStarted","Data":"dec4e9479380b7e735d9738838d634ee2790d7b2a87d040b08f5a5fd61367c2e"} Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.304520 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5pdcr"] Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.313381 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" event={"ID":"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb","Type":"ContainerStarted","Data":"2f124547d6c6b6e4cadb63b6c1732b81d7643f21b62f22762426569eabbfc7c4"} Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.315514 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.319572 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rz6g5"] Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.319624 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mthhs"] Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.320575 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-2z46g" event={"ID":"a79bbac6-f40a-4c92-8854-7ab5e72573cc","Type":"ContainerStarted","Data":"20eb94421e0eea64be64321e6c00b000dee79fd4a4662f6a369022dd15653f7e"} Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.324887 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-dzxd5" event={"ID":"768c9343-1391-4221-b008-5dee2921953f","Type":"ContainerStarted","Data":"b4f28a77963b985e8d55e24993d011cf12ef9fe6081a370032418a9d8ea0ffda"} Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.325582 4775 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-r6c6z container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.325618 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" podUID="f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.329001 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" event={"ID":"8bb7828f-6d99-4539-8312-c8e96bfbc608","Type":"ContainerStarted","Data":"ac8f4456eebf6c010d00843db2a2f64d8e826e8bf72978a78df02bc12cb5b96f"} Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.349772 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:04 crc kubenswrapper[4775]: E0321 04:51:04.349986 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:04.849961108 +0000 UTC m=+217.826424732 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.350194 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:04 crc kubenswrapper[4775]: E0321 04:51:04.351371 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:04.851363599 +0000 UTC m=+217.827827223 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.453464 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:04 crc kubenswrapper[4775]: E0321 04:51:04.454650 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:04.954633034 +0000 UTC m=+217.931096668 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:04 crc kubenswrapper[4775]: W0321 04:51:04.484652 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12b1cc27_6a60_43d2_9d5e_eb7a54c1e899.slice/crio-a0ddf7f265c9468b0e817c2b6ad7ddc35b80f659e965df1ca89854dfd66fbfcb WatchSource:0}: Error finding container a0ddf7f265c9468b0e817c2b6ad7ddc35b80f659e965df1ca89854dfd66fbfcb: Status 404 returned error can't find the container with id a0ddf7f265c9468b0e817c2b6ad7ddc35b80f659e965df1ca89854dfd66fbfcb Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.554497 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:04 crc kubenswrapper[4775]: E0321 04:51:04.554817 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:05.054805981 +0000 UTC m=+218.031269605 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.696108 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:04 crc kubenswrapper[4775]: E0321 04:51:04.696366 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:05.196344146 +0000 UTC m=+218.172807770 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.696888 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:04 crc kubenswrapper[4775]: E0321 04:51:04.697239 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:05.197224371 +0000 UTC m=+218.173688005 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.804812 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:04 crc kubenswrapper[4775]: E0321 04:51:04.805344 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:05.305322896 +0000 UTC m=+218.281786520 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:04 crc kubenswrapper[4775]: I0321 04:51:04.905855 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:04 crc kubenswrapper[4775]: E0321 04:51:04.906229 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:05.406216823 +0000 UTC m=+218.382680447 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:04.997592 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-nstnr" podStartSLOduration=158.997572017 podStartE2EDuration="2m38.997572017s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:04.971509219 +0000 UTC m=+217.947972843" watchObservedRunningTime="2026-03-21 04:51:04.997572017 +0000 UTC m=+217.974035641" Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:04.997814 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k5dks"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.001170 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jd86b"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.006632 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:05 crc kubenswrapper[4775]: E0321 04:51:05.007318 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:05.507298986 +0000 UTC m=+218.483762610 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.012265 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp"] Mar 21 04:51:05 crc kubenswrapper[4775]: W0321 04:51:05.017328 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded9baf4c_70b2_450f_9b21_76dfafbc44d0.slice/crio-1dc3c475bbed8c4530a54779f813884a49b5a06477a18a344d8eb396adfd545f WatchSource:0}: Error finding container 1dc3c475bbed8c4530a54779f813884a49b5a06477a18a344d8eb396adfd545f: Status 404 returned error can't find the container with id 1dc3c475bbed8c4530a54779f813884a49b5a06477a18a344d8eb396adfd545f Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.020183 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-wst2s"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.023740 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9v866"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.040780 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xv7nw"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.047546 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vjv4q"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.108316 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:05 crc kubenswrapper[4775]: E0321 04:51:05.108759 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:05.60874371 +0000 UTC m=+218.585207334 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.180808 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v296h"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.210241 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:05 crc kubenswrapper[4775]: E0321 04:51:05.210633 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:05.710613065 +0000 UTC m=+218.687076689 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.215920 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-725jh"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.218595 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5ztsd"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.227495 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-57cxh"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.235437 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7sfkh"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.244632 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zh8p"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.274614 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567810-wb89g"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.274667 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-zs5pr"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.278993 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wqjnj"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.283631 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-gd4gc"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.287061 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wrzs4"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.293173 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-d7xpn"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.295099 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mbgvc"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.296777 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-4nfs6"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.298495 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-hvfsp"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.312292 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:05 crc kubenswrapper[4775]: E0321 04:51:05.312694 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:05.812670606 +0000 UTC m=+218.789134230 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.346012 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5pdcr" event={"ID":"2d2512b8-9fb6-4b9b-9ac2-268e0b8bd04a","Type":"ContainerStarted","Data":"4ccdd23e8be4d6c9539771eef6195bf56b8b0ee2c73a8319460776e9e587d6d9"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.346291 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5pdcr" event={"ID":"2d2512b8-9fb6-4b9b-9ac2-268e0b8bd04a","Type":"ContainerStarted","Data":"1f7c118357de68461245eae47ffb437a1633b5749182b21c0c72fdfc4c679eb2"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.347832 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-468g4" event={"ID":"6343dd4f-5c5b-4c94-a4d6-f603698ba6ac","Type":"ContainerStarted","Data":"354589bc74d0e1df42a39518ab6dbad16d637dd99da7e74ffac93a01ae49e39c"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.350862 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wst2s" event={"ID":"a9ed1e0e-eff9-4690-bcf5-45f6074c200e","Type":"ContainerStarted","Data":"c582e2fab200954a9843b0514b416a8988bf80f007355b323c81a87c8930c6fa"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.351548 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" event={"ID":"fddde3da-8512-4e62-9c38-b59f98e117e0","Type":"ContainerStarted","Data":"bdc921caf55fdc9483dd16d86e33b3865a46d3e8a029b5b72e32afe0666cef10"} Mar 21 04:51:05 crc kubenswrapper[4775]: W0321 04:51:05.352323 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod918078e1_af38_475a_86d5_8179cafa18db.slice/crio-b1b80ce740efd38d0818a1758ffcc9a25de23ff4e3d9a46b3ed20d9fb4f1765f WatchSource:0}: Error finding container b1b80ce740efd38d0818a1758ffcc9a25de23ff4e3d9a46b3ed20d9fb4f1765f: Status 404 returned error can't find the container with id b1b80ce740efd38d0818a1758ffcc9a25de23ff4e3d9a46b3ed20d9fb4f1765f Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.352480 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-dzxd5" event={"ID":"768c9343-1391-4221-b008-5dee2921953f","Type":"ContainerStarted","Data":"378f686eba4695c392992cbcc55f4bb7d43751a0aa1f6827e445289cf7ab59ba"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.354673 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7sfkh" event={"ID":"2dbc150e-3a25-4b44-b01c-effe99de5152","Type":"ContainerStarted","Data":"81cfc7994200911020eefe9ee8577746708ae643101c25b2f8cec803d2f18463"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.355473 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5ztsd" event={"ID":"f9395e24-0d4b-4165-bf60-068876927f58","Type":"ContainerStarted","Data":"2f0781c3740928ce2d7eb2feb485c88f02d9ea7d5356a38d011e5a61cb72029a"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.356153 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vjv4q" event={"ID":"811f064c-ebf4-48ad-87a0-83205eb1eca5","Type":"ContainerStarted","Data":"c17997ba96f183ce91c800c29ded59676be28e8d3a7e31d33b00efb6bade2a36"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.357083 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" event={"ID":"892d15b6-460e-4892-a836-0cc284c8a326","Type":"ContainerStarted","Data":"16776553b2840887617e187a152ecb08b84ab564361957b0471e5c74a0b97f4f"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.357133 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" event={"ID":"892d15b6-460e-4892-a836-0cc284c8a326","Type":"ContainerStarted","Data":"575843aeea1bfa101dadda9f7c658bc97392556098e7a54921dab614fe073b20"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.359482 4775 generic.go:334] "Generic (PLEG): container finished" podID="8bb7828f-6d99-4539-8312-c8e96bfbc608" containerID="9377ec80f0257a109a3ef8a2ab177b965b751362a7bed153c19bdb616f962c5b" exitCode=0 Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.359550 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" event={"ID":"8bb7828f-6d99-4539-8312-c8e96bfbc608","Type":"ContainerDied","Data":"9377ec80f0257a109a3ef8a2ab177b965b751362a7bed153c19bdb616f962c5b"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.362066 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9v866" event={"ID":"10a9b5d9-308c-4971-a073-9c88de98d8ea","Type":"ContainerStarted","Data":"d391c9939f47b2c46b89dc19c1bfd9d652cf6bd95d985bf56a1c9557acbcc9fb"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.366828 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v296h" event={"ID":"b0b465ff-f8f3-4a99-9235-1ddc4ce093e4","Type":"ContainerStarted","Data":"006a4227f3b9a53c19aee02bd2ba8f8065dbb5adc9267be25a07167778e43098"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.367877 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mbgvc" event={"ID":"bc2b8485-8cc1-4029-8318-397d4278e455","Type":"ContainerStarted","Data":"0f6819a8fe627dd9cd2c2a54f184a37a7e20e8bcef4d909c3ed46add017746b7"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.369274 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" event={"ID":"ed9baf4c-70b2-450f-9b21-76dfafbc44d0","Type":"ContainerStarted","Data":"1dc3c475bbed8c4530a54779f813884a49b5a06477a18a344d8eb396adfd545f"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.370373 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-5rnj6" podStartSLOduration=159.370359623 podStartE2EDuration="2m39.370359623s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:05.368194341 +0000 UTC m=+218.344657975" watchObservedRunningTime="2026-03-21 04:51:05.370359623 +0000 UTC m=+218.346823247" Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.371424 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-725jh" event={"ID":"eddac2a3-8b07-410d-9e9f-f7ce0c05abf5","Type":"ContainerStarted","Data":"fac4dd9f3b6b0094e30d1053e0f38da29cf7c45cc03525a6d3a097acc02a041e"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.376541 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-fc4z8" event={"ID":"12b1cc27-6a60-43d2-9d5e-eb7a54c1e899","Type":"ContainerStarted","Data":"6f0ed735e075d381eb3244b0026ac1423e2945f8047e0e69ac0bbbe2cd398281"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.376589 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-fc4z8" event={"ID":"12b1cc27-6a60-43d2-9d5e-eb7a54c1e899","Type":"ContainerStarted","Data":"a0ddf7f265c9468b0e817c2b6ad7ddc35b80f659e965df1ca89854dfd66fbfcb"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.381096 4775 generic.go:334] "Generic (PLEG): container finished" podID="6e22b5c7-4191-4f21-82ba-3014ccc4e978" containerID="a5a2690a7310f199883212a85e22e16730b42ff31cd6f309adf16033e7c667f2" exitCode=0 Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.381191 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" event={"ID":"6e22b5c7-4191-4f21-82ba-3014ccc4e978","Type":"ContainerDied","Data":"a5a2690a7310f199883212a85e22e16730b42ff31cd6f309adf16033e7c667f2"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.381222 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" event={"ID":"6e22b5c7-4191-4f21-82ba-3014ccc4e978","Type":"ContainerStarted","Data":"0d54baee9468506d654bdead2edbe067a0624cb9f3d099cb47b6df832ca221c2"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.383558 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.384534 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-gd4gc" event={"ID":"1d080332-c215-44c1-a027-65afbc612f88","Type":"ContainerStarted","Data":"9efb482e4d2b0037c0e6660d1e9de43b853428582153c6ff1d5456d965e77cd6"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.385586 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-dnv7h" event={"ID":"ecae58ef-adaa-42ea-9246-49845334a819","Type":"ContainerStarted","Data":"e174b1b5351adfc3ac99f9c2b2eb0b849a681af6c4e663fff96b061851393e99"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.385611 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-dnv7h" event={"ID":"ecae58ef-adaa-42ea-9246-49845334a819","Type":"ContainerStarted","Data":"60c1166ebc7144e4d0ca3bd792fa45f2a6a46e6b7dd9e914c1abbb31f09ef797"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.386376 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xv7nw" event={"ID":"edb67e99-0d50-46e9-adbb-c0831dd915d8","Type":"ContainerStarted","Data":"a0107ab6b2aa658877064abe6b5ae6a91c533d203e88950d607e2469e04edc54"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.387038 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-57cxh" event={"ID":"68d61e65-8275-4862-9dae-a75029889b2a","Type":"ContainerStarted","Data":"7c38f9475d7e5c854a00a5490720d86e44d20915c3af5d6a9fb49e6ae7a38696"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.402092 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-2z46g" event={"ID":"a79bbac6-f40a-4c92-8854-7ab5e72573cc","Type":"ContainerStarted","Data":"a99acbaab0f76819308ab208aae43a9cd832350fb9839c3fa49ffcc6ee292566"} Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.402212 4775 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-r6c6z container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.402245 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" podUID="f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.403342 4775 patch_prober.go:28] interesting pod/downloads-7954f5f757-nstnr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.403374 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nstnr" podUID="c3d58eba-4ddf-463c-baa1-1943fb60c732" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 21 04:51:05 crc kubenswrapper[4775]: E0321 04:51:05.414145 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:05.914077058 +0000 UTC m=+218.890540672 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.414215 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.415028 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.415189 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 04:51:05 crc kubenswrapper[4775]: E0321 04:51:05.415497 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:05.915488369 +0000 UTC m=+218.891951993 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:05 crc kubenswrapper[4775]: W0321 04:51:05.442096 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02b06614_31da_49d9_bc97_cf61b065d42f.slice/crio-72a6014e8a8c00b4814a9127976a071de2c161ae17ce9251e2050d85143f64b4 WatchSource:0}: Error finding container 72a6014e8a8c00b4814a9127976a071de2c161ae17ce9251e2050d85143f64b4: Status 404 returned error can't find the container with id 72a6014e8a8c00b4814a9127976a071de2c161ae17ce9251e2050d85143f64b4 Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.451468 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" podStartSLOduration=159.451445371 podStartE2EDuration="2m39.451445371s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:05.450406302 +0000 UTC m=+218.426869926" watchObservedRunningTime="2026-03-21 04:51:05.451445371 +0000 UTC m=+218.427909005" Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.516519 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:05 crc kubenswrapper[4775]: E0321 04:51:05.518087 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:06.018062885 +0000 UTC m=+218.994526549 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.559993 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.590342 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.618934 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:05 crc kubenswrapper[4775]: E0321 04:51:05.619638 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:06.119619601 +0000 UTC m=+219.096083245 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.626594 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.636522 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zkl7p"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.653857 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-6jnn5"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.670486 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-blwmr"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.699005 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2lnkp"] Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.719837 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:05 crc kubenswrapper[4775]: E0321 04:51:05.720240 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:06.22022331 +0000 UTC m=+219.196686934 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:05 crc kubenswrapper[4775]: W0321 04:51:05.726671 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0bb62e0_24b1_4ee5_b8c0_077a8f2aeeb8.slice/crio-7a3e51d71262c4421a06e5c99d2e5a2ee3a4cb2aaf4dd59819720b466894bbfd WatchSource:0}: Error finding container 7a3e51d71262c4421a06e5c99d2e5a2ee3a4cb2aaf4dd59819720b466894bbfd: Status 404 returned error can't find the container with id 7a3e51d71262c4421a06e5c99d2e5a2ee3a4cb2aaf4dd59819720b466894bbfd Mar 21 04:51:05 crc kubenswrapper[4775]: W0321 04:51:05.732420 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5aa6958_e573_4efb_a031_218c62b0bec9.slice/crio-4a3a49e3ddeff3a00ab9609e9ae3b8c22b4ec782261392142e9e4f564705d42d WatchSource:0}: Error finding container 4a3a49e3ddeff3a00ab9609e9ae3b8c22b4ec782261392142e9e4f564705d42d: Status 404 returned error can't find the container with id 4a3a49e3ddeff3a00ab9609e9ae3b8c22b4ec782261392142e9e4f564705d42d Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.816869 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.818884 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.818930 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.821544 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:05 crc kubenswrapper[4775]: E0321 04:51:05.821901 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:06.3218829 +0000 UTC m=+219.298346524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.863487 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-dzxd5" podStartSLOduration=159.863468084 podStartE2EDuration="2m39.863468084s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:05.857524774 +0000 UTC m=+218.833988398" watchObservedRunningTime="2026-03-21 04:51:05.863468084 +0000 UTC m=+218.839931708" Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.922480 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:05 crc kubenswrapper[4775]: E0321 04:51:05.922624 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:06.422608223 +0000 UTC m=+219.399071847 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.922880 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:05 crc kubenswrapper[4775]: E0321 04:51:05.923158 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:06.423150368 +0000 UTC m=+219.399613992 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.931728 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-2z46g" podStartSLOduration=159.931707564 podStartE2EDuration="2m39.931707564s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:05.930130619 +0000 UTC m=+218.906594263" watchObservedRunningTime="2026-03-21 04:51:05.931707564 +0000 UTC m=+218.908171188" Mar 21 04:51:05 crc kubenswrapper[4775]: I0321 04:51:05.932366 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-468g4" podStartSLOduration=159.932359233 podStartE2EDuration="2m39.932359233s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:05.891900791 +0000 UTC m=+218.868364415" watchObservedRunningTime="2026-03-21 04:51:05.932359233 +0000 UTC m=+218.908822857" Mar 21 04:51:05 crc kubenswrapper[4775]: W0321 04:51:05.974648 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd55e67e9_9980_4878_9875_a07207894f6f.slice/crio-973fb4f2cc44ee4346ef8c713d7373ffea3fdbff448ecd56e7c8391ee6ce48e6 WatchSource:0}: Error finding container 973fb4f2cc44ee4346ef8c713d7373ffea3fdbff448ecd56e7c8391ee6ce48e6: Status 404 returned error can't find the container with id 973fb4f2cc44ee4346ef8c713d7373ffea3fdbff448ecd56e7c8391ee6ce48e6 Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.023916 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:06 crc kubenswrapper[4775]: E0321 04:51:06.024056 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:06.524030116 +0000 UTC m=+219.500493740 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.024213 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:06 crc kubenswrapper[4775]: E0321 04:51:06.024610 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:06.524587582 +0000 UTC m=+219.501051226 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.125323 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:06 crc kubenswrapper[4775]: E0321 04:51:06.125540 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:06.62551566 +0000 UTC m=+219.601979284 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.125920 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:06 crc kubenswrapper[4775]: E0321 04:51:06.126183 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:06.626173719 +0000 UTC m=+219.602637343 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.226460 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:06 crc kubenswrapper[4775]: E0321 04:51:06.226637 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:06.726613523 +0000 UTC m=+219.703077147 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.226759 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:06 crc kubenswrapper[4775]: E0321 04:51:06.227012 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:06.727001035 +0000 UTC m=+219.703464659 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.327598 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:06 crc kubenswrapper[4775]: E0321 04:51:06.327696 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:06.827678206 +0000 UTC m=+219.804141830 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.327785 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:06 crc kubenswrapper[4775]: E0321 04:51:06.328039 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:06.828030626 +0000 UTC m=+219.804494240 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.407762 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567810-wb89g" event={"ID":"cd25c8a4-8047-4602-a95b-3308af65bd38","Type":"ContainerStarted","Data":"05b88dca4e9352028f202490cd0cd91a11f14d92891caca0835ae23e4321056a"} Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.409352 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zs5pr" event={"ID":"ae87529c-e52f-45b0-9bbf-2a652e628bc5","Type":"ContainerStarted","Data":"67f32c8a2a75b425e13a6954ee446b7d5f82b03031a619ecb3e2549b702adedc"} Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.411063 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wqjnj" event={"ID":"918078e1-af38-475a-86d5-8179cafa18db","Type":"ContainerStarted","Data":"b1b80ce740efd38d0818a1758ffcc9a25de23ff4e3d9a46b3ed20d9fb4f1765f"} Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.413009 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zh8p" event={"ID":"1017f8be-2192-46a4-8717-ead73ca5e81b","Type":"ContainerStarted","Data":"14694a169bb5c4955d641d259ab6b1e11c221e9de14ac903c5c305b569f98734"} Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.414366 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" event={"ID":"dfbaac71-f99c-4373-a469-f2e5dd0ee632","Type":"ContainerStarted","Data":"73bfd1fd04d6bcc9e68b01d3bd224e4608501bca9349188673dd3d6681965588"} Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.415650 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct" event={"ID":"a5aa6958-e573-4efb-a031-218c62b0bec9","Type":"ContainerStarted","Data":"4a3a49e3ddeff3a00ab9609e9ae3b8c22b4ec782261392142e9e4f564705d42d"} Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.417366 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" event={"ID":"9de83f0b-7dd2-4846-a1ce-c8af930778f4","Type":"ContainerStarted","Data":"2731fba8805ee077aab9635f83ab8d0734ef47f44e0bcd4695582de53fd7acba"} Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.418335 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-d7xpn" event={"ID":"e5224539-6d29-4bc3-9656-4665eb287e28","Type":"ContainerStarted","Data":"631844f264e2c2c60c82cbb55b5b9b825b4c2c41854e332cffa0322b4537e9e1"} Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.419380 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" event={"ID":"07720715-40fb-4f7d-8f6b-381dc7cf9ea2","Type":"ContainerStarted","Data":"e203b5acaed5b8428d41bc14f17c02edfa1433e80d67ab20e5c1162cf529a5c5"} Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.420590 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2lnkp" event={"ID":"b0a24bc9-b395-4732-ba57-b096ee9ffdb1","Type":"ContainerStarted","Data":"5630ce022553fe34ac22d879a625a130a1d5f8464075a4e9b7f945e66d1d43db"} Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.421971 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" event={"ID":"a49709c7-59e0-440e-89c2-177c42cd28e8","Type":"ContainerStarted","Data":"7e4b00575887da7acfd9d6e07116fb0ebf6717dff53758a618d5c9d92741f392"} Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.426332 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-blwmr" event={"ID":"0c0dc3b9-8e11-4ff0-8f26-549ce47215f0","Type":"ContainerStarted","Data":"d60caa6b7c834cd2c81448a4d8ff279ec53ea7cd58de54b2a9f6ffeef78f88dd"} Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.427599 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zkl7p" event={"ID":"d55e67e9-9980-4878-9875-a07207894f6f","Type":"ContainerStarted","Data":"973fb4f2cc44ee4346ef8c713d7373ffea3fdbff448ecd56e7c8391ee6ce48e6"} Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.428619 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:06 crc kubenswrapper[4775]: E0321 04:51:06.429227 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:06.929201322 +0000 UTC m=+219.905664986 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.429894 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z" event={"ID":"f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8","Type":"ContainerStarted","Data":"7a3e51d71262c4421a06e5c99d2e5a2ee3a4cb2aaf4dd59819720b466894bbfd"} Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.432056 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-6jnn5" event={"ID":"35c6eec5-7117-48de-a4c5-640517da6a76","Type":"ContainerStarted","Data":"d4b3eac7e2e58a362d63264ddf58256cf443c42ec6564a89e5121145388f21ab"} Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.433297 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj" event={"ID":"02b06614-31da-49d9-bc97-cf61b065d42f","Type":"ContainerStarted","Data":"72a6014e8a8c00b4814a9127976a071de2c161ae17ce9251e2050d85143f64b4"} Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.470556 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-dnv7h" podStartSLOduration=6.470533359 podStartE2EDuration="6.470533359s" podCreationTimestamp="2026-03-21 04:51:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:06.469095057 +0000 UTC m=+219.445558691" watchObservedRunningTime="2026-03-21 04:51:06.470533359 +0000 UTC m=+219.446997003" Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.531539 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:06 crc kubenswrapper[4775]: E0321 04:51:06.533898 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:07.033871368 +0000 UTC m=+220.010335012 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.632553 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:06 crc kubenswrapper[4775]: E0321 04:51:06.632852 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:07.13283602 +0000 UTC m=+220.109299634 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.734290 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:06 crc kubenswrapper[4775]: E0321 04:51:06.734656 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:07.234635053 +0000 UTC m=+220.211098737 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.817817 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.817898 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.836095 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:06 crc kubenswrapper[4775]: E0321 04:51:06.836699 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:07.336646443 +0000 UTC m=+220.313110067 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.838973 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:06 crc kubenswrapper[4775]: E0321 04:51:06.841483 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:07.341453101 +0000 UTC m=+220.317916725 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:06 crc kubenswrapper[4775]: I0321 04:51:06.940412 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:06 crc kubenswrapper[4775]: E0321 04:51:06.940809 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:07.440787564 +0000 UTC m=+220.417251188 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.041861 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:07 crc kubenswrapper[4775]: E0321 04:51:07.042307 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:07.542286079 +0000 UTC m=+220.518749703 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.143787 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:07 crc kubenswrapper[4775]: E0321 04:51:07.144239 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:07.644224706 +0000 UTC m=+220.620688330 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.244876 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:07 crc kubenswrapper[4775]: E0321 04:51:07.245299 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:07.745284379 +0000 UTC m=+220.721748003 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.345513 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:07 crc kubenswrapper[4775]: E0321 04:51:07.345700 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:07.845677162 +0000 UTC m=+220.822140786 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.345831 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:07 crc kubenswrapper[4775]: E0321 04:51:07.346195 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:07.846183916 +0000 UTC m=+220.822647540 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.438790 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" event={"ID":"ed9baf4c-70b2-450f-9b21-76dfafbc44d0","Type":"ContainerStarted","Data":"cadd789c9a6b6b6dccb1af86d4359d6025e42837584cd6624c1767f1cecac34c"} Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.439934 4775 generic.go:334] "Generic (PLEG): container finished" podID="892d15b6-460e-4892-a836-0cc284c8a326" containerID="16776553b2840887617e187a152ecb08b84ab564361957b0471e5c74a0b97f4f" exitCode=0 Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.439989 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" event={"ID":"892d15b6-460e-4892-a836-0cc284c8a326","Type":"ContainerDied","Data":"16776553b2840887617e187a152ecb08b84ab564361957b0471e5c74a0b97f4f"} Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.441152 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k5dks" event={"ID":"edb4245a-7971-4c40-81b6-27d56b319a2f","Type":"ContainerStarted","Data":"93335dc6f90290b36d46a46319c144e689ee8335d4f0c54f8222753aeee4eb14"} Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.446727 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:07 crc kubenswrapper[4775]: E0321 04:51:07.447724 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:07.947705382 +0000 UTC m=+220.924168996 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.451858 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9v866" event={"ID":"10a9b5d9-308c-4971-a073-9c88de98d8ea","Type":"ContainerStarted","Data":"c99542331ad097e2f95ddfa1e7556cebdd46fb352aef4933c487cf230490583e"} Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.453628 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xv7nw" event={"ID":"edb67e99-0d50-46e9-adbb-c0831dd915d8","Type":"ContainerStarted","Data":"f3a29dd8f77e6a4b94910c0c0c74c2afe3c0b75e0fdd9a8fcf71d4f7770d5653"} Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.453789 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-fc4z8" Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.459795 4775 patch_prober.go:28] interesting pod/console-operator-58897d9998-fc4z8 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.459869 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-fc4z8" podUID="12b1cc27-6a60-43d2-9d5e-eb7a54c1e899" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.470901 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-fc4z8" podStartSLOduration=161.470886928 podStartE2EDuration="2m41.470886928s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:07.469065805 +0000 UTC m=+220.445529429" watchObservedRunningTime="2026-03-21 04:51:07.470886928 +0000 UTC m=+220.447350552" Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.484530 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5pdcr" podStartSLOduration=161.484513379 podStartE2EDuration="2m41.484513379s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:07.484285242 +0000 UTC m=+220.460748886" watchObservedRunningTime="2026-03-21 04:51:07.484513379 +0000 UTC m=+220.460977003" Mar 21 04:51:07 crc kubenswrapper[4775]: E0321 04:51:07.550465 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:08.050412722 +0000 UTC m=+221.026876346 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.549753 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.652465 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:07 crc kubenswrapper[4775]: E0321 04:51:07.652934 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:08.152912565 +0000 UTC m=+221.129376189 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.753783 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:07 crc kubenswrapper[4775]: E0321 04:51:07.754072 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:08.254060519 +0000 UTC m=+221.230524143 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.818484 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.818577 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.855200 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:07 crc kubenswrapper[4775]: E0321 04:51:07.855456 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:08.35542881 +0000 UTC m=+221.331892434 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.855681 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:07 crc kubenswrapper[4775]: E0321 04:51:07.855948 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:08.355941035 +0000 UTC m=+221.332404659 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.957357 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:07 crc kubenswrapper[4775]: E0321 04:51:07.957646 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:08.457608795 +0000 UTC m=+221.434072419 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:07 crc kubenswrapper[4775]: I0321 04:51:07.957755 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:07 crc kubenswrapper[4775]: E0321 04:51:07.958436 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:08.458424608 +0000 UTC m=+221.434888412 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.059834 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:08 crc kubenswrapper[4775]: E0321 04:51:08.059993 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:08.559963044 +0000 UTC m=+221.536426668 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.060302 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:08 crc kubenswrapper[4775]: E0321 04:51:08.060727 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:08.560717136 +0000 UTC m=+221.537180930 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.162420 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:08 crc kubenswrapper[4775]: E0321 04:51:08.162585 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:08.662554081 +0000 UTC m=+221.639017705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.162805 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:08 crc kubenswrapper[4775]: E0321 04:51:08.163274 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:08.663266561 +0000 UTC m=+221.639730175 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.264694 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:08 crc kubenswrapper[4775]: E0321 04:51:08.264901 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:08.764870689 +0000 UTC m=+221.741334313 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.265069 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:08 crc kubenswrapper[4775]: E0321 04:51:08.265430 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:08.765420685 +0000 UTC m=+221.741884309 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.368629 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:08 crc kubenswrapper[4775]: E0321 04:51:08.369018 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:08.86899978 +0000 UTC m=+221.845463404 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.460433 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v296h" event={"ID":"b0b465ff-f8f3-4a99-9235-1ddc4ce093e4","Type":"ContainerStarted","Data":"18c05d7f58a99612dc5efe5ad65e1a4f010d0505662765c2b5f2cf9c83622c20"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.462044 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-725jh" event={"ID":"eddac2a3-8b07-410d-9e9f-f7ce0c05abf5","Type":"ContainerStarted","Data":"9242d5c32349be04a03c56f0b3585e84f1337e2ab99ece891d25c14234eae8d5"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.463705 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" event={"ID":"8bb7828f-6d99-4539-8312-c8e96bfbc608","Type":"ContainerStarted","Data":"590ca6e3c97d44acdce1b2b5e4b30315b8b8022f7cd8a1624210df48524edcf8"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.466568 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2lnkp" event={"ID":"b0a24bc9-b395-4732-ba57-b096ee9ffdb1","Type":"ContainerStarted","Data":"a312561e8254adf11616d36ba39133cb34b7dc608bc2da8feb55af6233140c65"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.468001 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj" event={"ID":"02b06614-31da-49d9-bc97-cf61b065d42f","Type":"ContainerStarted","Data":"cbc98787ecc0c6c4832727fda1f187cf8a193cad3b2d5a9a0ebc67dae7ea78ea"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.469194 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k5dks" event={"ID":"edb4245a-7971-4c40-81b6-27d56b319a2f","Type":"ContainerStarted","Data":"9691da391ada20ada3b71bb2d7e26ca8ba0f89f4dce37b834b05ee3c9e082918"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.469655 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:08 crc kubenswrapper[4775]: E0321 04:51:08.469976 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:08.969963159 +0000 UTC m=+221.946426783 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.470686 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wqjnj" event={"ID":"918078e1-af38-475a-86d5-8179cafa18db","Type":"ContainerStarted","Data":"d43ab3ec1001720c6f472efab2a3f71284b8cb8e3357689550716f3a74fd014e"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.472331 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" event={"ID":"fddde3da-8512-4e62-9c38-b59f98e117e0","Type":"ContainerStarted","Data":"e7db5d2af2536f63217dcf1801711995965b013eb5a3eabb8f70b3e311b18d76"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.473825 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-gd4gc" event={"ID":"1d080332-c215-44c1-a027-65afbc612f88","Type":"ContainerStarted","Data":"0d2a7ce003a897c99bcb1c1e9cbeaa9b5ad53f73f02e92d0cc9aaa2361fe518d"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.475189 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vjv4q" event={"ID":"811f064c-ebf4-48ad-87a0-83205eb1eca5","Type":"ContainerStarted","Data":"91b9d97978afcef7d317d0be26aeb6112ab84e003138a0924c602ef35b20e3c5"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.476271 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-blwmr" event={"ID":"0c0dc3b9-8e11-4ff0-8f26-549ce47215f0","Type":"ContainerStarted","Data":"ecf2efe54a36b3fb6058b5bfe0d55546e6f3136149f8eabafeb20ba7213485d9"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.477402 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zkl7p" event={"ID":"d55e67e9-9980-4878-9875-a07207894f6f","Type":"ContainerStarted","Data":"3d338a745024ac75c81bf2abc4d2ef739c654f6537a34419b6ab721180950f6d"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.478498 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mbgvc" event={"ID":"bc2b8485-8cc1-4029-8318-397d4278e455","Type":"ContainerStarted","Data":"9edced4268caed98927ad877592a67e465613569cffdc221ea601b6eae827973"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.479807 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5ztsd" event={"ID":"f9395e24-0d4b-4165-bf60-068876927f58","Type":"ContainerStarted","Data":"25df534758847bc5ccc6009699a59047abe78191ef225b4f013fadcba9be1f14"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.481531 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" event={"ID":"dfbaac71-f99c-4373-a469-f2e5dd0ee632","Type":"ContainerStarted","Data":"e6613d34aa689f2c33b46ca7fb2fff89fd26737df8f8eb26a52c2030909ae41f"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.484326 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7sfkh" event={"ID":"2dbc150e-3a25-4b44-b01c-effe99de5152","Type":"ContainerStarted","Data":"5623ba27f455dcdbf40b96e8e00a12d1541682430e1b057b616d215082ae7128"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.488995 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct" event={"ID":"a5aa6958-e573-4efb-a031-218c62b0bec9","Type":"ContainerStarted","Data":"1b93f4361b64d716502128254e9ea25b78ba9ed772325baf5c04d50a5a77ab40"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.490925 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-d7xpn" event={"ID":"e5224539-6d29-4bc3-9656-4665eb287e28","Type":"ContainerStarted","Data":"75b230198fb6a70ca3ef179e91486c0ac56c1b4e6cb9288659a34229d0e5d109"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.492249 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-57cxh" event={"ID":"68d61e65-8275-4862-9dae-a75029889b2a","Type":"ContainerStarted","Data":"0fab95b37a2e6ed33a86e59e397871dbda6fae2eaae9335694addbaa577daf2b"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.493409 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" event={"ID":"a49709c7-59e0-440e-89c2-177c42cd28e8","Type":"ContainerStarted","Data":"8d28f7922639a7fd05cf65cac943f2930631c08e489a583966597a6f307142d7"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.494500 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zs5pr" event={"ID":"ae87529c-e52f-45b0-9bbf-2a652e628bc5","Type":"ContainerStarted","Data":"e2ab3430d2b9e7725b9fe97cfaf6740b73f817210e59fff19d1b2eb8d4004cce"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.495627 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-6jnn5" event={"ID":"35c6eec5-7117-48de-a4c5-640517da6a76","Type":"ContainerStarted","Data":"8b385a6616a741e413b81fd1ad0ea83118f452733c49d9fa7d2ed79f401bb413"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.508042 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z" event={"ID":"f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8","Type":"ContainerStarted","Data":"e12324fff0bd481ca62c439bc0330f51748fe72d84ac697cfe3adde23279b170"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.510663 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" event={"ID":"07720715-40fb-4f7d-8f6b-381dc7cf9ea2","Type":"ContainerStarted","Data":"4b385123fb6a15d90b1cec5c84204e7c26de4b5079a1b283ad23d8b734086ae6"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.512329 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" event={"ID":"892d15b6-460e-4892-a836-0cc284c8a326","Type":"ContainerStarted","Data":"849182efddb8c3413751c9d3506283faddbabe30d171beba88cbea21627385e0"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.513725 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wst2s" event={"ID":"a9ed1e0e-eff9-4690-bcf5-45f6074c200e","Type":"ContainerStarted","Data":"4843fbb3995c43ae7854ec367e570829c36ae201e54901d87c5c625dcdce3764"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.515597 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zh8p" event={"ID":"1017f8be-2192-46a4-8717-ead73ca5e81b","Type":"ContainerStarted","Data":"af2a3164a7ad4571de890915ec43f2b5bddc2dae35ae42a0b0637e6a7790886d"} Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.517351 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.517383 4775 patch_prober.go:28] interesting pod/console-operator-58897d9998-fc4z8 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.517416 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-fc4z8" podUID="12b1cc27-6a60-43d2-9d5e-eb7a54c1e899" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.517949 4775 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-55lcp container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.517979 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" podUID="ed9baf4c-70b2-450f-9b21-76dfafbc44d0" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.541219 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xv7nw" podStartSLOduration=162.541186064 podStartE2EDuration="2m42.541186064s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:08.53892791 +0000 UTC m=+221.515391534" watchObservedRunningTime="2026-03-21 04:51:08.541186064 +0000 UTC m=+221.517649688" Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.576616 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:08 crc kubenswrapper[4775]: E0321 04:51:08.577884 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:09.077864508 +0000 UTC m=+222.054328132 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.619100 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" podStartSLOduration=162.619078251 podStartE2EDuration="2m42.619078251s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:08.599717255 +0000 UTC m=+221.576180879" watchObservedRunningTime="2026-03-21 04:51:08.619078251 +0000 UTC m=+221.595541875" Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.620019 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9v866" podStartSLOduration=162.620012448 podStartE2EDuration="2m42.620012448s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:08.617344482 +0000 UTC m=+221.593808116" watchObservedRunningTime="2026-03-21 04:51:08.620012448 +0000 UTC m=+221.596476082" Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.678234 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:08 crc kubenswrapper[4775]: E0321 04:51:08.678684 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:09.178667743 +0000 UTC m=+222.155131367 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.779528 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:08 crc kubenswrapper[4775]: E0321 04:51:08.779678 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:09.279659553 +0000 UTC m=+222.256123177 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.780139 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:08 crc kubenswrapper[4775]: E0321 04:51:08.780413 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:09.280405204 +0000 UTC m=+222.256868828 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.839639 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:08 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:08 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:08 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.839697 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.881740 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:08 crc kubenswrapper[4775]: E0321 04:51:08.882088 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:09.382073694 +0000 UTC m=+222.358537318 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:08 crc kubenswrapper[4775]: I0321 04:51:08.983676 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:08 crc kubenswrapper[4775]: E0321 04:51:08.983965 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:09.48395441 +0000 UTC m=+222.460418024 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.084891 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:09 crc kubenswrapper[4775]: E0321 04:51:09.085052 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:09.585021403 +0000 UTC m=+222.561485027 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.085132 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:09 crc kubenswrapper[4775]: E0321 04:51:09.085439 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:09.585428304 +0000 UTC m=+222.561891928 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.186536 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:09 crc kubenswrapper[4775]: E0321 04:51:09.186731 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:09.686705983 +0000 UTC m=+222.663169607 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.186783 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:09 crc kubenswrapper[4775]: E0321 04:51:09.187169 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:09.687156866 +0000 UTC m=+222.663620490 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.287965 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:09 crc kubenswrapper[4775]: E0321 04:51:09.288198 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:09.788176987 +0000 UTC m=+222.764640611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.288310 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:09 crc kubenswrapper[4775]: E0321 04:51:09.288667 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:09.788658011 +0000 UTC m=+222.765121635 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.388973 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:09 crc kubenswrapper[4775]: E0321 04:51:09.389143 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:09.889094535 +0000 UTC m=+222.865558159 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.389285 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:09 crc kubenswrapper[4775]: E0321 04:51:09.389716 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:09.889679592 +0000 UTC m=+222.866143216 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.492793 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:09 crc kubenswrapper[4775]: E0321 04:51:09.493366 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:09.993340559 +0000 UTC m=+222.969804193 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.520773 4775 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-55lcp container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.520844 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" podUID="ed9baf4c-70b2-450f-9b21-76dfafbc44d0" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.522044 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.522236 4775 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-jd86b container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" start-of-body= Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.522280 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" podUID="fddde3da-8512-4e62-9c38-b59f98e117e0" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.544354 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" podStartSLOduration=163.544338904 podStartE2EDuration="2m43.544338904s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:09.542671386 +0000 UTC m=+222.519135000" watchObservedRunningTime="2026-03-21 04:51:09.544338904 +0000 UTC m=+222.520802528" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.558498 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zkl7p" podStartSLOduration=163.55847487 podStartE2EDuration="2m43.55847487s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:09.557740609 +0000 UTC m=+222.534204243" watchObservedRunningTime="2026-03-21 04:51:09.55847487 +0000 UTC m=+222.534938504" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.591782 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" podStartSLOduration=163.591764096 podStartE2EDuration="2m43.591764096s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:09.590561951 +0000 UTC m=+222.567025575" watchObservedRunningTime="2026-03-21 04:51:09.591764096 +0000 UTC m=+222.568227720" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.594752 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:09 crc kubenswrapper[4775]: E0321 04:51:09.612903 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:10.112877822 +0000 UTC m=+223.089341446 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.622446 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v296h" podStartSLOduration=163.622425346 podStartE2EDuration="2m43.622425346s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:09.618634198 +0000 UTC m=+222.595097842" watchObservedRunningTime="2026-03-21 04:51:09.622425346 +0000 UTC m=+222.598888970" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.639505 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-4nfs6" podStartSLOduration=163.639488926 podStartE2EDuration="2m43.639488926s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:09.638536049 +0000 UTC m=+222.614999673" watchObservedRunningTime="2026-03-21 04:51:09.639488926 +0000 UTC m=+222.615952550" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.696651 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:09 crc kubenswrapper[4775]: E0321 04:51:09.696809 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:10.196778862 +0000 UTC m=+223.173242486 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.697190 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:09 crc kubenswrapper[4775]: E0321 04:51:09.697433 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:10.19742288 +0000 UTC m=+223.173886504 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.707325 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" podStartSLOduration=163.707307634 podStartE2EDuration="2m43.707307634s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:09.65841997 +0000 UTC m=+222.634883584" watchObservedRunningTime="2026-03-21 04:51:09.707307634 +0000 UTC m=+222.683771268" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.708539 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" podStartSLOduration=163.708531789 podStartE2EDuration="2m43.708531789s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:09.704586586 +0000 UTC m=+222.681050210" watchObservedRunningTime="2026-03-21 04:51:09.708531789 +0000 UTC m=+222.684995413" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.745204 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" podStartSLOduration=163.745185092 podStartE2EDuration="2m43.745185092s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:09.72492891 +0000 UTC m=+222.701392544" watchObservedRunningTime="2026-03-21 04:51:09.745185092 +0000 UTC m=+222.721648716" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.763076 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-gd4gc" podStartSLOduration=163.763056275 podStartE2EDuration="2m43.763056275s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:09.762352885 +0000 UTC m=+222.738816509" watchObservedRunningTime="2026-03-21 04:51:09.763056275 +0000 UTC m=+222.739519909" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.763453 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-6jnn5" podStartSLOduration=9.763446106 podStartE2EDuration="9.763446106s" podCreationTimestamp="2026-03-21 04:51:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:09.748560139 +0000 UTC m=+222.725023763" watchObservedRunningTime="2026-03-21 04:51:09.763446106 +0000 UTC m=+222.739909730" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.796837 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5ztsd" podStartSLOduration=163.796819175 podStartE2EDuration="2m43.796819175s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:09.783290996 +0000 UTC m=+222.759754640" watchObservedRunningTime="2026-03-21 04:51:09.796819175 +0000 UTC m=+222.773282799" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.798254 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:09 crc kubenswrapper[4775]: E0321 04:51:09.798389 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:10.298372169 +0000 UTC m=+223.274835793 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.798650 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:09 crc kubenswrapper[4775]: E0321 04:51:09.798957 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:10.298946066 +0000 UTC m=+223.275409740 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.799558 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-725jh" podStartSLOduration=163.799548873 podStartE2EDuration="2m43.799548873s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:09.796631849 +0000 UTC m=+222.773095483" watchObservedRunningTime="2026-03-21 04:51:09.799548873 +0000 UTC m=+222.776012497" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.816724 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-blwmr" podStartSLOduration=163.816706336 podStartE2EDuration="2m43.816706336s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:09.812474414 +0000 UTC m=+222.788938048" watchObservedRunningTime="2026-03-21 04:51:09.816706336 +0000 UTC m=+222.793169960" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.819543 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:09 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:09 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:09 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.819595 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.828417 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wqjnj" podStartSLOduration=163.828401372 podStartE2EDuration="2m43.828401372s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:09.828312459 +0000 UTC m=+222.804776083" watchObservedRunningTime="2026-03-21 04:51:09.828401372 +0000 UTC m=+222.804864996" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.856975 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct" podStartSLOduration=163.856957682 podStartE2EDuration="2m43.856957682s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:09.855763178 +0000 UTC m=+222.832226802" watchObservedRunningTime="2026-03-21 04:51:09.856957682 +0000 UTC m=+222.833421306" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.858053 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-d7xpn" podStartSLOduration=163.858046443 podStartE2EDuration="2m43.858046443s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:09.844315769 +0000 UTC m=+222.820779383" watchObservedRunningTime="2026-03-21 04:51:09.858046443 +0000 UTC m=+222.834510077" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.891089 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2zh8p" podStartSLOduration=163.891072892 podStartE2EDuration="2m43.891072892s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:09.874030452 +0000 UTC m=+222.850494076" watchObservedRunningTime="2026-03-21 04:51:09.891072892 +0000 UTC m=+222.867536516" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.892725 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-wst2s" podStartSLOduration=163.892719319 podStartE2EDuration="2m43.892719319s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:09.889345702 +0000 UTC m=+222.865809326" watchObservedRunningTime="2026-03-21 04:51:09.892719319 +0000 UTC m=+222.869182943" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.899639 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:09 crc kubenswrapper[4775]: E0321 04:51:09.899778 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:10.399759441 +0000 UTC m=+223.376223065 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.899890 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:09 crc kubenswrapper[4775]: E0321 04:51:09.900184 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:10.400177913 +0000 UTC m=+223.376641537 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.925298 4775 ???:1] "http: TLS handshake error from 192.168.126.11:43608: no serving certificate available for the kubelet" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.966030 4775 ???:1] "http: TLS handshake error from 192.168.126.11:43610: no serving certificate available for the kubelet" Mar 21 04:51:09 crc kubenswrapper[4775]: I0321 04:51:09.991979 4775 ???:1] "http: TLS handshake error from 192.168.126.11:43622: no serving certificate available for the kubelet" Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.001002 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:10 crc kubenswrapper[4775]: E0321 04:51:10.001201 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:10.501174904 +0000 UTC m=+223.477638528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.060806 4775 ???:1] "http: TLS handshake error from 192.168.126.11:43630: no serving certificate available for the kubelet" Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.079137 4775 ???:1] "http: TLS handshake error from 192.168.126.11:43640: no serving certificate available for the kubelet" Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.102427 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:10 crc kubenswrapper[4775]: E0321 04:51:10.102800 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:10.602782952 +0000 UTC m=+223.579246576 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.164100 4775 ???:1] "http: TLS handshake error from 192.168.126.11:43644: no serving certificate available for the kubelet" Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.203616 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:10 crc kubenswrapper[4775]: E0321 04:51:10.204050 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:10.704015729 +0000 UTC m=+223.680479393 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.305661 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:10 crc kubenswrapper[4775]: E0321 04:51:10.306069 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:10.806053569 +0000 UTC m=+223.782517193 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.351065 4775 ???:1] "http: TLS handshake error from 192.168.126.11:43648: no serving certificate available for the kubelet" Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.407324 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:10 crc kubenswrapper[4775]: E0321 04:51:10.407646 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:10.907595896 +0000 UTC m=+223.884059520 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.508998 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:10 crc kubenswrapper[4775]: E0321 04:51:10.509423 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:11.00940382 +0000 UTC m=+223.985867444 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.527923 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z" event={"ID":"f0bb62e0-24b1-4ee5-b8c0-077a8f2aeeb8","Type":"ContainerStarted","Data":"a8337c07a9f91d9d03dd6039d3198c7ccdfcdfb956af5d440f3c4e0bd2d75826"} Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.531156 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" event={"ID":"6e22b5c7-4191-4f21-82ba-3014ccc4e978","Type":"ContainerStarted","Data":"9d744d38f317071bd769b3e7072f8fab149a6bbf4171f825b3b0d5e6cdaf2588"} Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.536020 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zs5pr" event={"ID":"ae87529c-e52f-45b0-9bbf-2a652e628bc5","Type":"ContainerStarted","Data":"8ad0a1aa3f4ed910a583b987476e23f0c356f3beb7b29b09ef9f390ee28dfdcd"} Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.536158 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-zs5pr" Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.538044 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2lnkp" event={"ID":"b0a24bc9-b395-4732-ba57-b096ee9ffdb1","Type":"ContainerStarted","Data":"719ce0747f05487b39ed1cd009257ecf52b6ac8ce8de87d8bcfdf50d6d1e4082"} Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.543885 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj" event={"ID":"02b06614-31da-49d9-bc97-cf61b065d42f","Type":"ContainerStarted","Data":"defd41f64a00fac497123059d0e52b8cafbca23fb2bba9bf796fb94315df2e30"} Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.550205 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vsh8z" podStartSLOduration=164.550184291 podStartE2EDuration="2m44.550184291s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:10.54843468 +0000 UTC m=+223.524898324" watchObservedRunningTime="2026-03-21 04:51:10.550184291 +0000 UTC m=+223.526647915" Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.554437 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mbgvc" event={"ID":"bc2b8485-8cc1-4029-8318-397d4278e455","Type":"ContainerStarted","Data":"1e8daaf27e70ace4c3af569a43c51111922c374bd73158cd17848aaae6a12515"} Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.555070 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mbgvc" Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.562869 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-57cxh" event={"ID":"68d61e65-8275-4862-9dae-a75029889b2a","Type":"ContainerStarted","Data":"ac1cbe68dd73ea5c8caacc3b618dba0ce5e706de8e4108c1b4c621d135d2097c"} Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.575955 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vjv4q" event={"ID":"811f064c-ebf4-48ad-87a0-83205eb1eca5","Type":"ContainerStarted","Data":"422091c765551c45cf53928bcb6bbf7cb726ca790f73aa24731ad99b2ab51f79"} Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.576069 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8kpqj" podStartSLOduration=164.576044923 podStartE2EDuration="2m44.576044923s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:10.571909935 +0000 UTC m=+223.548373559" watchObservedRunningTime="2026-03-21 04:51:10.576044923 +0000 UTC m=+223.552508547" Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.588323 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k5dks" event={"ID":"edb4245a-7971-4c40-81b6-27d56b319a2f","Type":"ContainerStarted","Data":"f0dec679b3b89379b4d0e57d8a3c9e1005a582f726c8219969bad32fa27e3e7a"} Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.602356 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7sfkh" event={"ID":"2dbc150e-3a25-4b44-b01c-effe99de5152","Type":"ContainerStarted","Data":"c79b3392d38d1173c38edbc74e51d7e1ebb61bf920208dfdec4b3116dbff3597"} Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.603108 4775 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-jd86b container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" start-of-body= Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.603176 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" podUID="fddde3da-8512-4e62-9c38-b59f98e117e0" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.607751 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-2lnkp" podStartSLOduration=164.607734333 podStartE2EDuration="2m44.607734333s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:10.607176277 +0000 UTC m=+223.583639911" watchObservedRunningTime="2026-03-21 04:51:10.607734333 +0000 UTC m=+223.584197947" Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.610243 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:10 crc kubenswrapper[4775]: E0321 04:51:10.610348 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:11.110325938 +0000 UTC m=+224.086789562 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.611457 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:10 crc kubenswrapper[4775]: E0321 04:51:10.613344 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:11.113332014 +0000 UTC m=+224.089795638 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.660894 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7sfkh" podStartSLOduration=164.66087484 podStartE2EDuration="2m44.66087484s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:10.659709236 +0000 UTC m=+223.636172860" watchObservedRunningTime="2026-03-21 04:51:10.66087484 +0000 UTC m=+223.637338464" Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.662246 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-zs5pr" podStartSLOduration=10.662237269 podStartE2EDuration="10.662237269s" podCreationTimestamp="2026-03-21 04:51:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:10.643922493 +0000 UTC m=+223.620386117" watchObservedRunningTime="2026-03-21 04:51:10.662237269 +0000 UTC m=+223.638700893" Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.684264 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mbgvc" podStartSLOduration=164.684241251 podStartE2EDuration="2m44.684241251s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:10.683547311 +0000 UTC m=+223.660010935" watchObservedRunningTime="2026-03-21 04:51:10.684241251 +0000 UTC m=+223.660704875" Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.697010 4775 ???:1] "http: TLS handshake error from 192.168.126.11:34200: no serving certificate available for the kubelet" Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.702399 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-57cxh" podStartSLOduration=164.702378942 podStartE2EDuration="2m44.702378942s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:10.701534417 +0000 UTC m=+223.677998041" watchObservedRunningTime="2026-03-21 04:51:10.702378942 +0000 UTC m=+223.678842566" Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.712758 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:10 crc kubenswrapper[4775]: E0321 04:51:10.714690 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:11.214675545 +0000 UTC m=+224.191139169 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.740125 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-vjv4q" podStartSLOduration=164.739965931 podStartE2EDuration="2m44.739965931s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:10.736862802 +0000 UTC m=+223.713326436" watchObservedRunningTime="2026-03-21 04:51:10.739965931 +0000 UTC m=+223.716429555" Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.765906 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k5dks" podStartSLOduration=164.765872565 podStartE2EDuration="2m44.765872565s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:10.763797045 +0000 UTC m=+223.740260669" watchObservedRunningTime="2026-03-21 04:51:10.765872565 +0000 UTC m=+223.742336189" Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.817958 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:10 crc kubenswrapper[4775]: E0321 04:51:10.818290 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:11.31827768 +0000 UTC m=+224.294741294 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.822609 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:10 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:10 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:10 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.822643 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:10 crc kubenswrapper[4775]: I0321 04:51:10.918948 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:10 crc kubenswrapper[4775]: E0321 04:51:10.919327 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:11.419313152 +0000 UTC m=+224.395776776 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:11 crc kubenswrapper[4775]: I0321 04:51:11.020186 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:11 crc kubenswrapper[4775]: E0321 04:51:11.020510 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:11.520497308 +0000 UTC m=+224.496960932 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:11 crc kubenswrapper[4775]: I0321 04:51:11.121067 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:11 crc kubenswrapper[4775]: E0321 04:51:11.121256 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:11.62122932 +0000 UTC m=+224.597692944 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:11 crc kubenswrapper[4775]: I0321 04:51:11.121335 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:11 crc kubenswrapper[4775]: E0321 04:51:11.121662 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:11.621653173 +0000 UTC m=+224.598116797 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:11 crc kubenswrapper[4775]: I0321 04:51:11.222564 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:11 crc kubenswrapper[4775]: E0321 04:51:11.222763 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:11.722737216 +0000 UTC m=+224.699200850 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:11 crc kubenswrapper[4775]: I0321 04:51:11.223075 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:11 crc kubenswrapper[4775]: E0321 04:51:11.223451 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:11.723441266 +0000 UTC m=+224.699904880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:11 crc kubenswrapper[4775]: I0321 04:51:11.324134 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:11 crc kubenswrapper[4775]: E0321 04:51:11.324576 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:11.824555429 +0000 UTC m=+224.801019053 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:11 crc kubenswrapper[4775]: I0321 04:51:11.390791 4775 ???:1] "http: TLS handshake error from 192.168.126.11:34206: no serving certificate available for the kubelet" Mar 21 04:51:11 crc kubenswrapper[4775]: I0321 04:51:11.426295 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:11 crc kubenswrapper[4775]: E0321 04:51:11.426609 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:11.926593589 +0000 UTC m=+224.903057213 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:11 crc kubenswrapper[4775]: I0321 04:51:11.527408 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:11 crc kubenswrapper[4775]: E0321 04:51:11.528191 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:12.028169156 +0000 UTC m=+225.004632780 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:11 crc kubenswrapper[4775]: I0321 04:51:11.621443 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" event={"ID":"6e22b5c7-4191-4f21-82ba-3014ccc4e978","Type":"ContainerStarted","Data":"ea1b2f10b0c4b51db9b95325c0c76b92a590974e1b40882e8f0862257853dcab"} Mar 21 04:51:11 crc kubenswrapper[4775]: I0321 04:51:11.627460 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" event={"ID":"9de83f0b-7dd2-4846-a1ce-c8af930778f4","Type":"ContainerStarted","Data":"c74b513c5d5a5ec41643b4b209c53ff6b64295bf34309a4ab65491776584358e"} Mar 21 04:51:11 crc kubenswrapper[4775]: I0321 04:51:11.630073 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:11 crc kubenswrapper[4775]: E0321 04:51:11.630420 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:12.130406712 +0000 UTC m=+225.106870326 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:11 crc kubenswrapper[4775]: I0321 04:51:11.678209 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" podStartSLOduration=165.678191085 podStartE2EDuration="2m45.678191085s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:11.677260368 +0000 UTC m=+224.653724002" watchObservedRunningTime="2026-03-21 04:51:11.678191085 +0000 UTC m=+224.654654709" Mar 21 04:51:11 crc kubenswrapper[4775]: I0321 04:51:11.731466 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:11 crc kubenswrapper[4775]: E0321 04:51:11.732971 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:12.232952067 +0000 UTC m=+225.209415691 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:11 crc kubenswrapper[4775]: I0321 04:51:11.826756 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:11 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:11 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:11 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:11 crc kubenswrapper[4775]: I0321 04:51:11.826815 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:11 crc kubenswrapper[4775]: I0321 04:51:11.833728 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:11 crc kubenswrapper[4775]: E0321 04:51:11.834194 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:12.334173344 +0000 UTC m=+225.310636968 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:11 crc kubenswrapper[4775]: I0321 04:51:11.935542 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:11 crc kubenswrapper[4775]: E0321 04:51:11.935730 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:12.43570224 +0000 UTC m=+225.412165864 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:11 crc kubenswrapper[4775]: I0321 04:51:11.935886 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:11 crc kubenswrapper[4775]: E0321 04:51:11.936215 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:12.436206065 +0000 UTC m=+225.412669689 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.036961 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:12 crc kubenswrapper[4775]: E0321 04:51:12.037097 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:12.537080342 +0000 UTC m=+225.513543956 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.037290 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:12 crc kubenswrapper[4775]: E0321 04:51:12.037608 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:12.537597957 +0000 UTC m=+225.514061591 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.138883 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:12 crc kubenswrapper[4775]: E0321 04:51:12.139056 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:12.63902916 +0000 UTC m=+225.615492784 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.139213 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:12 crc kubenswrapper[4775]: E0321 04:51:12.139524 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:12.639514994 +0000 UTC m=+225.615978618 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.168779 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.170895 4775 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-mthhs container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.170934 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" podUID="892d15b6-460e-4892-a836-0cc284c8a326" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.170898 4775 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-mthhs container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.171081 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" podUID="892d15b6-460e-4892-a836-0cc284c8a326" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.171290 4775 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-mthhs container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.171325 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" podUID="892d15b6-460e-4892-a836-0cc284c8a326" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.240239 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:12 crc kubenswrapper[4775]: E0321 04:51:12.240399 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:12.74037579 +0000 UTC m=+225.716839414 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.240534 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:12 crc kubenswrapper[4775]: E0321 04:51:12.240764 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:12.740756651 +0000 UTC m=+225.717220275 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.341517 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:12 crc kubenswrapper[4775]: E0321 04:51:12.341839 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:12.841825904 +0000 UTC m=+225.818289528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.443227 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:12 crc kubenswrapper[4775]: E0321 04:51:12.443549 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:12.943538065 +0000 UTC m=+225.920001689 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.544079 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:12 crc kubenswrapper[4775]: E0321 04:51:12.544502 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:13.044469553 +0000 UTC m=+226.020933177 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.544851 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:12 crc kubenswrapper[4775]: E0321 04:51:12.545224 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:13.045208885 +0000 UTC m=+226.021672529 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.645649 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:12 crc kubenswrapper[4775]: E0321 04:51:12.646071 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:13.146050991 +0000 UTC m=+226.122514625 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.743125 4775 ???:1] "http: TLS handshake error from 192.168.126.11:34214: no serving certificate available for the kubelet" Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.747157 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:12 crc kubenswrapper[4775]: E0321 04:51:12.748366 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:13.248348949 +0000 UTC m=+226.224812573 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.819737 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:12 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:12 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:12 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.819809 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.848187 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:12 crc kubenswrapper[4775]: E0321 04:51:12.848610 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:13.348592988 +0000 UTC m=+226.325056612 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.950328 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:12 crc kubenswrapper[4775]: E0321 04:51:12.950745 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:13.450725481 +0000 UTC m=+226.427189105 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:12 crc kubenswrapper[4775]: I0321 04:51:12.956025 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.023916 4775 patch_prober.go:28] interesting pod/downloads-7954f5f757-nstnr container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.023982 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-nstnr" podUID="c3d58eba-4ddf-463c-baa1-1943fb60c732" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.024349 4775 patch_prober.go:28] interesting pod/downloads-7954f5f757-nstnr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.024404 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nstnr" podUID="c3d58eba-4ddf-463c-baa1-1943fb60c732" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.051873 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:13 crc kubenswrapper[4775]: E0321 04:51:13.052083 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:13.552052021 +0000 UTC m=+226.528515645 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.052338 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:13 crc kubenswrapper[4775]: E0321 04:51:13.052716 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:13.55270489 +0000 UTC m=+226.529168524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.154071 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:13 crc kubenswrapper[4775]: E0321 04:51:13.154415 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:13.65438887 +0000 UTC m=+226.630852514 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.154560 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:13 crc kubenswrapper[4775]: E0321 04:51:13.154852 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:13.654845313 +0000 UTC m=+226.631308937 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.157745 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.157849 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.255997 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:13 crc kubenswrapper[4775]: E0321 04:51:13.256392 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:13.756365978 +0000 UTC m=+226.732829602 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.256536 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:13 crc kubenswrapper[4775]: E0321 04:51:13.256893 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:13.756875503 +0000 UTC m=+226.733339127 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.258673 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.258707 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.259788 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.260539 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.261625 4775 patch_prober.go:28] interesting pod/console-f9d7485db-wst2s container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.261656 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-wst2s" podUID="a9ed1e0e-eff9-4690-bcf5-45f6074c200e" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.261634 4775 patch_prober.go:28] interesting pod/apiserver-76f77b778f-rz6g5 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.27:8443/livez\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.261727 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" podUID="6e22b5c7-4191-4f21-82ba-3014ccc4e978" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.27:8443/livez\": dial tcp 10.217.0.27:8443: connect: connection refused" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.266901 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-fc4z8" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.308040 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.357413 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:13 crc kubenswrapper[4775]: E0321 04:51:13.358746 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:13.858724528 +0000 UTC m=+226.835188152 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.433781 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.435619 4775 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wrzs4 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.435684 4775 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wrzs4 container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.435910 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" podUID="dfbaac71-f99c-4373-a469-f2e5dd0ee632" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.435806 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" podUID="dfbaac71-f99c-4373-a469-f2e5dd0ee632" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.436479 4775 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wrzs4 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.436569 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" podUID="dfbaac71-f99c-4373-a469-f2e5dd0ee632" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.459227 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:13 crc kubenswrapper[4775]: E0321 04:51:13.459678 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:13.959661327 +0000 UTC m=+226.936124951 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.471336 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.560378 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:13 crc kubenswrapper[4775]: E0321 04:51:13.561352 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:14.061319736 +0000 UTC m=+227.037783360 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.647874 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" event={"ID":"9de83f0b-7dd2-4846-a1ce-c8af930778f4","Type":"ContainerStarted","Data":"dac9469a3446348249e3c427c296e28a0c4f474cc224cec4d3ca45da6d0d9051"} Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.661825 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:13 crc kubenswrapper[4775]: E0321 04:51:13.662147 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:14.162135732 +0000 UTC m=+227.138599356 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.713627 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v296h" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.762724 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:13 crc kubenswrapper[4775]: E0321 04:51:13.764451 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:14.26443127 +0000 UTC m=+227.240894884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.774711 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v296h" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.810154 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.812774 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.813069 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.815422 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.815732 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.816422 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.824990 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:13 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:13 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:13 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.825371 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.826486 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.866673 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:13 crc kubenswrapper[4775]: E0321 04:51:13.868725 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:14.368710264 +0000 UTC m=+227.345173898 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.906467 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.920433 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kxtlf" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.971425 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:13 crc kubenswrapper[4775]: E0321 04:51:13.971589 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:14.471562428 +0000 UTC m=+227.448026052 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.973304 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f1cf896-5a0b-43e5-b36b-ba705e02825f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0f1cf896-5a0b-43e5-b36b-ba705e02825f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.974076 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:13 crc kubenswrapper[4775]: I0321 04:51:13.974271 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0f1cf896-5a0b-43e5-b36b-ba705e02825f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0f1cf896-5a0b-43e5-b36b-ba705e02825f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 21 04:51:13 crc kubenswrapper[4775]: E0321 04:51:13.976752 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:14.476734827 +0000 UTC m=+227.453198451 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.008564 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-959kj"] Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.009572 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-959kj" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.017076 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.053242 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-959kj"] Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.075881 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.076194 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f1cf896-5a0b-43e5-b36b-ba705e02825f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0f1cf896-5a0b-43e5-b36b-ba705e02825f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.076276 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0f1cf896-5a0b-43e5-b36b-ba705e02825f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0f1cf896-5a0b-43e5-b36b-ba705e02825f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.076341 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0f1cf896-5a0b-43e5-b36b-ba705e02825f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0f1cf896-5a0b-43e5-b36b-ba705e02825f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 21 04:51:14 crc kubenswrapper[4775]: E0321 04:51:14.076409 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:14.576395609 +0000 UTC m=+227.552859233 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.133183 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f1cf896-5a0b-43e5-b36b-ba705e02825f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0f1cf896-5a0b-43e5-b36b-ba705e02825f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.133451 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.178854 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bb65dee-cd5f-46b3-9e7d-36e5d182d19e-utilities\") pod \"certified-operators-959kj\" (UID: \"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e\") " pod="openshift-marketplace/certified-operators-959kj" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.179147 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.179180 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ww6h\" (UniqueName: \"kubernetes.io/projected/5bb65dee-cd5f-46b3-9e7d-36e5d182d19e-kube-api-access-7ww6h\") pod \"certified-operators-959kj\" (UID: \"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e\") " pod="openshift-marketplace/certified-operators-959kj" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.179205 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bb65dee-cd5f-46b3-9e7d-36e5d182d19e-catalog-content\") pod \"certified-operators-959kj\" (UID: \"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e\") " pod="openshift-marketplace/certified-operators-959kj" Mar 21 04:51:14 crc kubenswrapper[4775]: E0321 04:51:14.179478 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:14.679466159 +0000 UTC m=+227.655929783 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.200028 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wbtgv"] Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.201011 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wbtgv" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.205387 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.238899 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wbtgv"] Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.258355 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zkl7p" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.281646 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zkl7p" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.281859 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:14 crc kubenswrapper[4775]: E0321 04:51:14.281971 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:14.781944292 +0000 UTC m=+227.758407916 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.282237 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.282266 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccb910ab-dcef-4523-81df-c0fb5eb83429-catalog-content\") pod \"community-operators-wbtgv\" (UID: \"ccb910ab-dcef-4523-81df-c0fb5eb83429\") " pod="openshift-marketplace/community-operators-wbtgv" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.282293 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ww6h\" (UniqueName: \"kubernetes.io/projected/5bb65dee-cd5f-46b3-9e7d-36e5d182d19e-kube-api-access-7ww6h\") pod \"certified-operators-959kj\" (UID: \"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e\") " pod="openshift-marketplace/certified-operators-959kj" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.282314 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92fzw\" (UniqueName: \"kubernetes.io/projected/ccb910ab-dcef-4523-81df-c0fb5eb83429-kube-api-access-92fzw\") pod \"community-operators-wbtgv\" (UID: \"ccb910ab-dcef-4523-81df-c0fb5eb83429\") " pod="openshift-marketplace/community-operators-wbtgv" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.282338 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bb65dee-cd5f-46b3-9e7d-36e5d182d19e-catalog-content\") pod \"certified-operators-959kj\" (UID: \"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e\") " pod="openshift-marketplace/certified-operators-959kj" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.282355 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccb910ab-dcef-4523-81df-c0fb5eb83429-utilities\") pod \"community-operators-wbtgv\" (UID: \"ccb910ab-dcef-4523-81df-c0fb5eb83429\") " pod="openshift-marketplace/community-operators-wbtgv" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.282397 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bb65dee-cd5f-46b3-9e7d-36e5d182d19e-utilities\") pod \"certified-operators-959kj\" (UID: \"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e\") " pod="openshift-marketplace/certified-operators-959kj" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.282800 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bb65dee-cd5f-46b3-9e7d-36e5d182d19e-catalog-content\") pod \"certified-operators-959kj\" (UID: \"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e\") " pod="openshift-marketplace/certified-operators-959kj" Mar 21 04:51:14 crc kubenswrapper[4775]: E0321 04:51:14.282934 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:14.78291995 +0000 UTC m=+227.759383634 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.283034 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bb65dee-cd5f-46b3-9e7d-36e5d182d19e-utilities\") pod \"certified-operators-959kj\" (UID: \"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e\") " pod="openshift-marketplace/certified-operators-959kj" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.353455 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ww6h\" (UniqueName: \"kubernetes.io/projected/5bb65dee-cd5f-46b3-9e7d-36e5d182d19e-kube-api-access-7ww6h\") pod \"certified-operators-959kj\" (UID: \"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e\") " pod="openshift-marketplace/certified-operators-959kj" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.383764 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.383959 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92fzw\" (UniqueName: \"kubernetes.io/projected/ccb910ab-dcef-4523-81df-c0fb5eb83429-kube-api-access-92fzw\") pod \"community-operators-wbtgv\" (UID: \"ccb910ab-dcef-4523-81df-c0fb5eb83429\") " pod="openshift-marketplace/community-operators-wbtgv" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.384014 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccb910ab-dcef-4523-81df-c0fb5eb83429-utilities\") pod \"community-operators-wbtgv\" (UID: \"ccb910ab-dcef-4523-81df-c0fb5eb83429\") " pod="openshift-marketplace/community-operators-wbtgv" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.384110 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccb910ab-dcef-4523-81df-c0fb5eb83429-catalog-content\") pod \"community-operators-wbtgv\" (UID: \"ccb910ab-dcef-4523-81df-c0fb5eb83429\") " pod="openshift-marketplace/community-operators-wbtgv" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.384827 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccb910ab-dcef-4523-81df-c0fb5eb83429-catalog-content\") pod \"community-operators-wbtgv\" (UID: \"ccb910ab-dcef-4523-81df-c0fb5eb83429\") " pod="openshift-marketplace/community-operators-wbtgv" Mar 21 04:51:14 crc kubenswrapper[4775]: E0321 04:51:14.384899 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:14.884884548 +0000 UTC m=+227.861348172 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.385845 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccb910ab-dcef-4523-81df-c0fb5eb83429-utilities\") pod \"community-operators-wbtgv\" (UID: \"ccb910ab-dcef-4523-81df-c0fb5eb83429\") " pod="openshift-marketplace/community-operators-wbtgv" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.395990 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v5fgf"] Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.415890 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v5fgf" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.426443 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v5fgf"] Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.445785 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92fzw\" (UniqueName: \"kubernetes.io/projected/ccb910ab-dcef-4523-81df-c0fb5eb83429-kube-api-access-92fzw\") pod \"community-operators-wbtgv\" (UID: \"ccb910ab-dcef-4523-81df-c0fb5eb83429\") " pod="openshift-marketplace/community-operators-wbtgv" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.485319 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.485407 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de54a48a-b733-4042-80b6-ecc719712314-catalog-content\") pod \"certified-operators-v5fgf\" (UID: \"de54a48a-b733-4042-80b6-ecc719712314\") " pod="openshift-marketplace/certified-operators-v5fgf" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.485432 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de54a48a-b733-4042-80b6-ecc719712314-utilities\") pod \"certified-operators-v5fgf\" (UID: \"de54a48a-b733-4042-80b6-ecc719712314\") " pod="openshift-marketplace/certified-operators-v5fgf" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.485464 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98dgw\" (UniqueName: \"kubernetes.io/projected/de54a48a-b733-4042-80b6-ecc719712314-kube-api-access-98dgw\") pod \"certified-operators-v5fgf\" (UID: \"de54a48a-b733-4042-80b6-ecc719712314\") " pod="openshift-marketplace/certified-operators-v5fgf" Mar 21 04:51:14 crc kubenswrapper[4775]: E0321 04:51:14.485672 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:14.985656112 +0000 UTC m=+227.962119736 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.540190 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wbtgv" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.589661 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.589824 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98dgw\" (UniqueName: \"kubernetes.io/projected/de54a48a-b733-4042-80b6-ecc719712314-kube-api-access-98dgw\") pod \"certified-operators-v5fgf\" (UID: \"de54a48a-b733-4042-80b6-ecc719712314\") " pod="openshift-marketplace/certified-operators-v5fgf" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.589913 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de54a48a-b733-4042-80b6-ecc719712314-catalog-content\") pod \"certified-operators-v5fgf\" (UID: \"de54a48a-b733-4042-80b6-ecc719712314\") " pod="openshift-marketplace/certified-operators-v5fgf" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.589943 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de54a48a-b733-4042-80b6-ecc719712314-utilities\") pod \"certified-operators-v5fgf\" (UID: \"de54a48a-b733-4042-80b6-ecc719712314\") " pod="openshift-marketplace/certified-operators-v5fgf" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.590372 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de54a48a-b733-4042-80b6-ecc719712314-utilities\") pod \"certified-operators-v5fgf\" (UID: \"de54a48a-b733-4042-80b6-ecc719712314\") " pod="openshift-marketplace/certified-operators-v5fgf" Mar 21 04:51:14 crc kubenswrapper[4775]: E0321 04:51:14.590446 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:15.090430671 +0000 UTC m=+228.066894295 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.591238 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de54a48a-b733-4042-80b6-ecc719712314-catalog-content\") pod \"certified-operators-v5fgf\" (UID: \"de54a48a-b733-4042-80b6-ecc719712314\") " pod="openshift-marketplace/certified-operators-v5fgf" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.621883 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-svf9w"] Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.623012 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svf9w" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.630359 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98dgw\" (UniqueName: \"kubernetes.io/projected/de54a48a-b733-4042-80b6-ecc719712314-kube-api-access-98dgw\") pod \"certified-operators-v5fgf\" (UID: \"de54a48a-b733-4042-80b6-ecc719712314\") " pod="openshift-marketplace/certified-operators-v5fgf" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.635596 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-svf9w"] Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.644229 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-959kj" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.645446 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.696947 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dcbca72-150a-47c6-ac3c-f701ae82e05b-catalog-content\") pod \"community-operators-svf9w\" (UID: \"1dcbca72-150a-47c6-ac3c-f701ae82e05b\") " pod="openshift-marketplace/community-operators-svf9w" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.697043 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.697068 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wq48\" (UniqueName: \"kubernetes.io/projected/1dcbca72-150a-47c6-ac3c-f701ae82e05b-kube-api-access-4wq48\") pod \"community-operators-svf9w\" (UID: \"1dcbca72-150a-47c6-ac3c-f701ae82e05b\") " pod="openshift-marketplace/community-operators-svf9w" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.697103 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dcbca72-150a-47c6-ac3c-f701ae82e05b-utilities\") pod \"community-operators-svf9w\" (UID: \"1dcbca72-150a-47c6-ac3c-f701ae82e05b\") " pod="openshift-marketplace/community-operators-svf9w" Mar 21 04:51:14 crc kubenswrapper[4775]: E0321 04:51:14.701501 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:15.201483461 +0000 UTC m=+228.177947085 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.754416 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" event={"ID":"9de83f0b-7dd2-4846-a1ce-c8af930778f4","Type":"ContainerStarted","Data":"262e7c1fa0b36c90633bf27e707ad1c5a6e35614948e91b3bd14fdfe9c8383ff"} Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.764627 4775 generic.go:334] "Generic (PLEG): container finished" podID="a5aa6958-e573-4efb-a031-218c62b0bec9" containerID="1b93f4361b64d716502128254e9ea25b78ba9ed772325baf5c04d50a5a77ab40" exitCode=0 Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.765853 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct" event={"ID":"a5aa6958-e573-4efb-a031-218c62b0bec9","Type":"ContainerDied","Data":"1b93f4361b64d716502128254e9ea25b78ba9ed772325baf5c04d50a5a77ab40"} Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.813263 4775 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-sp69c container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:5443/healthz\": context deadline exceeded" start-of-body= Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.813642 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" podUID="a49709c7-59e0-440e-89c2-177c42cd28e8" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.33:5443/healthz\": context deadline exceeded" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.813260 4775 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-sp69c container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.33:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.814049 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" podUID="a49709c7-59e0-440e-89c2-177c42cd28e8" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.33:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.814604 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.814855 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wq48\" (UniqueName: \"kubernetes.io/projected/1dcbca72-150a-47c6-ac3c-f701ae82e05b-kube-api-access-4wq48\") pod \"community-operators-svf9w\" (UID: \"1dcbca72-150a-47c6-ac3c-f701ae82e05b\") " pod="openshift-marketplace/community-operators-svf9w" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.814903 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dcbca72-150a-47c6-ac3c-f701ae82e05b-utilities\") pod \"community-operators-svf9w\" (UID: \"1dcbca72-150a-47c6-ac3c-f701ae82e05b\") " pod="openshift-marketplace/community-operators-svf9w" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.814945 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dcbca72-150a-47c6-ac3c-f701ae82e05b-catalog-content\") pod \"community-operators-svf9w\" (UID: \"1dcbca72-150a-47c6-ac3c-f701ae82e05b\") " pod="openshift-marketplace/community-operators-svf9w" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.815397 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dcbca72-150a-47c6-ac3c-f701ae82e05b-catalog-content\") pod \"community-operators-svf9w\" (UID: \"1dcbca72-150a-47c6-ac3c-f701ae82e05b\") " pod="openshift-marketplace/community-operators-svf9w" Mar 21 04:51:14 crc kubenswrapper[4775]: E0321 04:51:14.815464 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:15.315451084 +0000 UTC m=+228.291914708 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.815841 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dcbca72-150a-47c6-ac3c-f701ae82e05b-utilities\") pod \"community-operators-svf9w\" (UID: \"1dcbca72-150a-47c6-ac3c-f701ae82e05b\") " pod="openshift-marketplace/community-operators-svf9w" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.832605 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-sp69c" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.834024 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:14 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:14 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:14 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.834085 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.847598 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v5fgf" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.851899 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wq48\" (UniqueName: \"kubernetes.io/projected/1dcbca72-150a-47c6-ac3c-f701ae82e05b-kube-api-access-4wq48\") pod \"community-operators-svf9w\" (UID: \"1dcbca72-150a-47c6-ac3c-f701ae82e05b\") " pod="openshift-marketplace/community-operators-svf9w" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.922457 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:14 crc kubenswrapper[4775]: E0321 04:51:14.926786 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:15.426756299 +0000 UTC m=+228.403219923 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.940489 4775 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.944858 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r6c6z"] Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.945089 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" podUID="f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb" containerName="controller-manager" containerID="cri-o://2f124547d6c6b6e4cadb63b6c1732b81d7643f21b62f22762426569eabbfc7c4" gracePeriod=30 Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.960171 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svf9w" Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.986513 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp"] Mar 21 04:51:14 crc kubenswrapper[4775]: I0321 04:51:14.987008 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" podUID="ed9baf4c-70b2-450f-9b21-76dfafbc44d0" containerName="route-controller-manager" containerID="cri-o://cadd789c9a6b6b6dccb1af86d4359d6025e42837584cd6624c1767f1cecac34c" gracePeriod=30 Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.024357 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:15 crc kubenswrapper[4775]: E0321 04:51:15.024747 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:15.524730353 +0000 UTC m=+228.501193977 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.128864 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:15 crc kubenswrapper[4775]: E0321 04:51:15.129240 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:15.629226224 +0000 UTC m=+228.605689848 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.173161 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wbtgv"] Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.181512 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mthhs" Mar 21 04:51:15 crc kubenswrapper[4775]: E0321 04:51:15.230045 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:15.730010439 +0000 UTC m=+228.706474063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.230149 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.230777 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:15 crc kubenswrapper[4775]: E0321 04:51:15.231173 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:15.73110591 +0000 UTC m=+228.707569534 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.248128 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-959kj"] Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.331704 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:15 crc kubenswrapper[4775]: E0321 04:51:15.332778 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:15.832763739 +0000 UTC m=+228.809227363 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.335136 4775 ???:1] "http: TLS handshake error from 192.168.126.11:34228: no serving certificate available for the kubelet" Mar 21 04:51:15 crc kubenswrapper[4775]: W0321 04:51:15.383226 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5bb65dee_cd5f_46b3_9e7d_36e5d182d19e.slice/crio-68f05861e537811c2b45eab073d2762025cd79d47c769cec2072bbee0ee4a322 WatchSource:0}: Error finding container 68f05861e537811c2b45eab073d2762025cd79d47c769cec2072bbee0ee4a322: Status 404 returned error can't find the container with id 68f05861e537811c2b45eab073d2762025cd79d47c769cec2072bbee0ee4a322 Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.398348 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v5fgf"] Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.433463 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:15 crc kubenswrapper[4775]: E0321 04:51:15.433915 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:15.933895874 +0000 UTC m=+228.910359508 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.530461 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-svf9w"] Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.536775 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:15 crc kubenswrapper[4775]: E0321 04:51:15.537206 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:16.03719001 +0000 UTC m=+229.013653634 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.563282 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.564099 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.566358 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.566625 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.584964 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.638464 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:15 crc kubenswrapper[4775]: E0321 04:51:15.638730 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:16.138719666 +0000 UTC m=+229.115183290 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.739110 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:15 crc kubenswrapper[4775]: E0321 04:51:15.739342 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 04:51:16.239290125 +0000 UTC m=+229.215753749 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.739507 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/829c993f-4963-416a-8d66-15cf19a68237-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"829c993f-4963-416a-8d66-15cf19a68237\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.739588 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.739696 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/829c993f-4963-416a-8d66-15cf19a68237-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"829c993f-4963-416a-8d66-15cf19a68237\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 21 04:51:15 crc kubenswrapper[4775]: E0321 04:51:15.739907 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 04:51:16.239894712 +0000 UTC m=+229.216358336 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hvd8t" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.789978 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5fgf" event={"ID":"de54a48a-b733-4042-80b6-ecc719712314","Type":"ContainerStarted","Data":"ade6fa22510b087e9954c736cf8d4e9716ee7ec2fc194bbd254a73a9267ce630"} Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.791406 4775 generic.go:334] "Generic (PLEG): container finished" podID="f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb" containerID="2f124547d6c6b6e4cadb63b6c1732b81d7643f21b62f22762426569eabbfc7c4" exitCode=0 Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.791472 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" event={"ID":"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb","Type":"ContainerDied","Data":"2f124547d6c6b6e4cadb63b6c1732b81d7643f21b62f22762426569eabbfc7c4"} Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.792357 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svf9w" event={"ID":"1dcbca72-150a-47c6-ac3c-f701ae82e05b","Type":"ContainerStarted","Data":"e49106633d57c504f2a2b545d76a34559e4299f9b5c7c912c88ab2862f4e055a"} Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.793147 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbtgv" event={"ID":"ccb910ab-dcef-4523-81df-c0fb5eb83429","Type":"ContainerStarted","Data":"c9000a4269c650286aea4fb24ec488fdb1ea75dd0febdc2786cb6046e9da1f69"} Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.794770 4775 generic.go:334] "Generic (PLEG): container finished" podID="ed9baf4c-70b2-450f-9b21-76dfafbc44d0" containerID="cadd789c9a6b6b6dccb1af86d4359d6025e42837584cd6624c1767f1cecac34c" exitCode=0 Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.794813 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" event={"ID":"ed9baf4c-70b2-450f-9b21-76dfafbc44d0","Type":"ContainerDied","Data":"cadd789c9a6b6b6dccb1af86d4359d6025e42837584cd6624c1767f1cecac34c"} Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.795604 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-959kj" event={"ID":"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e","Type":"ContainerStarted","Data":"68f05861e537811c2b45eab073d2762025cd79d47c769cec2072bbee0ee4a322"} Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.796603 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0f1cf896-5a0b-43e5-b36b-ba705e02825f","Type":"ContainerStarted","Data":"9a39bc499dfd59cb84df2219ab5f515cedb58c236d866f34ae7e0567a7995091"} Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.813920 4775 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-21T04:51:14.940509444Z","Handler":null,"Name":""} Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.826646 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:15 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:15 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:15 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.826714 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.840240 4775 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.840281 4775 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.841559 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.841992 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/829c993f-4963-416a-8d66-15cf19a68237-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"829c993f-4963-416a-8d66-15cf19a68237\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.842104 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/829c993f-4963-416a-8d66-15cf19a68237-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"829c993f-4963-416a-8d66-15cf19a68237\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.842255 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/829c993f-4963-416a-8d66-15cf19a68237-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"829c993f-4963-416a-8d66-15cf19a68237\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.864366 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.874504 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/829c993f-4963-416a-8d66-15cf19a68237-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"829c993f-4963-416a-8d66-15cf19a68237\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.889811 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.943230 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.973214 4775 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 04:51:15 crc kubenswrapper[4775]: I0321 04:51:15.973271 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.040321 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hvd8t\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.174393 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 21 04:51:16 crc kubenswrapper[4775]: W0321 04:51:16.228102 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod829c993f_4963_416a_8d66_15cf19a68237.slice/crio-8ee6fc7e5443705634d1b844abee45703b8c3f989c11bb3335bb9bbcf4253920 WatchSource:0}: Error finding container 8ee6fc7e5443705634d1b844abee45703b8c3f989c11bb3335bb9bbcf4253920: Status 404 returned error can't find the container with id 8ee6fc7e5443705634d1b844abee45703b8c3f989c11bb3335bb9bbcf4253920 Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.230786 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.289674 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.296964 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.299284 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.347991 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wm9fk\" (UniqueName: \"kubernetes.io/projected/a5aa6958-e573-4efb-a031-218c62b0bec9-kube-api-access-wm9fk\") pod \"a5aa6958-e573-4efb-a031-218c62b0bec9\" (UID: \"a5aa6958-e573-4efb-a031-218c62b0bec9\") " Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.348084 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5aa6958-e573-4efb-a031-218c62b0bec9-config-volume\") pod \"a5aa6958-e573-4efb-a031-218c62b0bec9\" (UID: \"a5aa6958-e573-4efb-a031-218c62b0bec9\") " Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.348196 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5aa6958-e573-4efb-a031-218c62b0bec9-secret-volume\") pod \"a5aa6958-e573-4efb-a031-218c62b0bec9\" (UID: \"a5aa6958-e573-4efb-a031-218c62b0bec9\") " Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.350996 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5aa6958-e573-4efb-a031-218c62b0bec9-config-volume" (OuterVolumeSpecName: "config-volume") pod "a5aa6958-e573-4efb-a031-218c62b0bec9" (UID: "a5aa6958-e573-4efb-a031-218c62b0bec9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.354378 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5aa6958-e573-4efb-a031-218c62b0bec9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a5aa6958-e573-4efb-a031-218c62b0bec9" (UID: "a5aa6958-e573-4efb-a031-218c62b0bec9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.357514 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5aa6958-e573-4efb-a031-218c62b0bec9-kube-api-access-wm9fk" (OuterVolumeSpecName: "kube-api-access-wm9fk") pod "a5aa6958-e573-4efb-a031-218c62b0bec9" (UID: "a5aa6958-e573-4efb-a031-218c62b0bec9"). InnerVolumeSpecName "kube-api-access-wm9fk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.394893 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hldm7"] Mar 21 04:51:16 crc kubenswrapper[4775]: E0321 04:51:16.395202 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5aa6958-e573-4efb-a031-218c62b0bec9" containerName="collect-profiles" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.395216 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5aa6958-e573-4efb-a031-218c62b0bec9" containerName="collect-profiles" Mar 21 04:51:16 crc kubenswrapper[4775]: E0321 04:51:16.395232 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb" containerName="controller-manager" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.395238 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb" containerName="controller-manager" Mar 21 04:51:16 crc kubenswrapper[4775]: E0321 04:51:16.395248 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed9baf4c-70b2-450f-9b21-76dfafbc44d0" containerName="route-controller-manager" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.395255 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed9baf4c-70b2-450f-9b21-76dfafbc44d0" containerName="route-controller-manager" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.395347 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5aa6958-e573-4efb-a031-218c62b0bec9" containerName="collect-profiles" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.395365 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb" containerName="controller-manager" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.395376 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed9baf4c-70b2-450f-9b21-76dfafbc44d0" containerName="route-controller-manager" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.396035 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hldm7" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.397856 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.426014 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hldm7"] Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.449261 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-client-ca\") pod \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\" (UID: \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\") " Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.449304 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zb5h\" (UniqueName: \"kubernetes.io/projected/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-kube-api-access-2zb5h\") pod \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\" (UID: \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\") " Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.449347 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7h98\" (UniqueName: \"kubernetes.io/projected/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-kube-api-access-r7h98\") pod \"ed9baf4c-70b2-450f-9b21-76dfafbc44d0\" (UID: \"ed9baf4c-70b2-450f-9b21-76dfafbc44d0\") " Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.449384 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-proxy-ca-bundles\") pod \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\" (UID: \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\") " Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.449413 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-config\") pod \"ed9baf4c-70b2-450f-9b21-76dfafbc44d0\" (UID: \"ed9baf4c-70b2-450f-9b21-76dfafbc44d0\") " Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.449601 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-config\") pod \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\" (UID: \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\") " Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.449630 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-serving-cert\") pod \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\" (UID: \"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb\") " Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.449644 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-client-ca\") pod \"ed9baf4c-70b2-450f-9b21-76dfafbc44d0\" (UID: \"ed9baf4c-70b2-450f-9b21-76dfafbc44d0\") " Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.449682 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-serving-cert\") pod \"ed9baf4c-70b2-450f-9b21-76dfafbc44d0\" (UID: \"ed9baf4c-70b2-450f-9b21-76dfafbc44d0\") " Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.449915 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5aa6958-e573-4efb-a031-218c62b0bec9-config-volume\") on node \"crc\" DevicePath \"\"" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.449927 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5aa6958-e573-4efb-a031-218c62b0bec9-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.449937 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wm9fk\" (UniqueName: \"kubernetes.io/projected/a5aa6958-e573-4efb-a031-218c62b0bec9-kube-api-access-wm9fk\") on node \"crc\" DevicePath \"\"" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.450529 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-client-ca" (OuterVolumeSpecName: "client-ca") pod "ed9baf4c-70b2-450f-9b21-76dfafbc44d0" (UID: "ed9baf4c-70b2-450f-9b21-76dfafbc44d0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.450668 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-config" (OuterVolumeSpecName: "config") pod "ed9baf4c-70b2-450f-9b21-76dfafbc44d0" (UID: "ed9baf4c-70b2-450f-9b21-76dfafbc44d0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.451586 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-client-ca" (OuterVolumeSpecName: "client-ca") pod "f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb" (UID: "f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.452778 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-config" (OuterVolumeSpecName: "config") pod "f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb" (UID: "f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.452847 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb" (UID: "f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.454291 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ed9baf4c-70b2-450f-9b21-76dfafbc44d0" (UID: "ed9baf4c-70b2-450f-9b21-76dfafbc44d0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.454835 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-kube-api-access-r7h98" (OuterVolumeSpecName: "kube-api-access-r7h98") pod "ed9baf4c-70b2-450f-9b21-76dfafbc44d0" (UID: "ed9baf4c-70b2-450f-9b21-76dfafbc44d0"). InnerVolumeSpecName "kube-api-access-r7h98". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.454920 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb" (UID: "f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.456202 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-kube-api-access-2zb5h" (OuterVolumeSpecName: "kube-api-access-2zb5h") pod "f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb" (UID: "f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb"). InnerVolumeSpecName "kube-api-access-2zb5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.541862 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hvd8t"] Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.552005 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gdv6\" (UniqueName: \"kubernetes.io/projected/571e84f2-a2bc-4f09-ac53-d4a4adafa80b-kube-api-access-2gdv6\") pod \"redhat-marketplace-hldm7\" (UID: \"571e84f2-a2bc-4f09-ac53-d4a4adafa80b\") " pod="openshift-marketplace/redhat-marketplace-hldm7" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.552376 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/571e84f2-a2bc-4f09-ac53-d4a4adafa80b-utilities\") pod \"redhat-marketplace-hldm7\" (UID: \"571e84f2-a2bc-4f09-ac53-d4a4adafa80b\") " pod="openshift-marketplace/redhat-marketplace-hldm7" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.552432 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/571e84f2-a2bc-4f09-ac53-d4a4adafa80b-catalog-content\") pod \"redhat-marketplace-hldm7\" (UID: \"571e84f2-a2bc-4f09-ac53-d4a4adafa80b\") " pod="openshift-marketplace/redhat-marketplace-hldm7" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.552740 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.552766 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.552777 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.552788 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.552801 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zb5h\" (UniqueName: \"kubernetes.io/projected/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-kube-api-access-2zb5h\") on node \"crc\" DevicePath \"\"" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.552816 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7h98\" (UniqueName: \"kubernetes.io/projected/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-kube-api-access-r7h98\") on node \"crc\" DevicePath \"\"" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.552828 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.552841 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed9baf4c-70b2-450f-9b21-76dfafbc44d0-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.552946 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:51:16 crc kubenswrapper[4775]: W0321 04:51:16.573558 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60cd23b8_6e1c_492e_aeb5_8b16609d06d1.slice/crio-224b2bc46f2dabb71d4e0596c8a7cb1f7d3a4099e704fb5ab5882b52edcb5f8a WatchSource:0}: Error finding container 224b2bc46f2dabb71d4e0596c8a7cb1f7d3a4099e704fb5ab5882b52edcb5f8a: Status 404 returned error can't find the container with id 224b2bc46f2dabb71d4e0596c8a7cb1f7d3a4099e704fb5ab5882b52edcb5f8a Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.654487 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gdv6\" (UniqueName: \"kubernetes.io/projected/571e84f2-a2bc-4f09-ac53-d4a4adafa80b-kube-api-access-2gdv6\") pod \"redhat-marketplace-hldm7\" (UID: \"571e84f2-a2bc-4f09-ac53-d4a4adafa80b\") " pod="openshift-marketplace/redhat-marketplace-hldm7" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.655872 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/571e84f2-a2bc-4f09-ac53-d4a4adafa80b-utilities\") pod \"redhat-marketplace-hldm7\" (UID: \"571e84f2-a2bc-4f09-ac53-d4a4adafa80b\") " pod="openshift-marketplace/redhat-marketplace-hldm7" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.656388 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/571e84f2-a2bc-4f09-ac53-d4a4adafa80b-utilities\") pod \"redhat-marketplace-hldm7\" (UID: \"571e84f2-a2bc-4f09-ac53-d4a4adafa80b\") " pod="openshift-marketplace/redhat-marketplace-hldm7" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.656852 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/571e84f2-a2bc-4f09-ac53-d4a4adafa80b-catalog-content\") pod \"redhat-marketplace-hldm7\" (UID: \"571e84f2-a2bc-4f09-ac53-d4a4adafa80b\") " pod="openshift-marketplace/redhat-marketplace-hldm7" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.658319 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/571e84f2-a2bc-4f09-ac53-d4a4adafa80b-catalog-content\") pod \"redhat-marketplace-hldm7\" (UID: \"571e84f2-a2bc-4f09-ac53-d4a4adafa80b\") " pod="openshift-marketplace/redhat-marketplace-hldm7" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.676381 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gdv6\" (UniqueName: \"kubernetes.io/projected/571e84f2-a2bc-4f09-ac53-d4a4adafa80b-kube-api-access-2gdv6\") pod \"redhat-marketplace-hldm7\" (UID: \"571e84f2-a2bc-4f09-ac53-d4a4adafa80b\") " pod="openshift-marketplace/redhat-marketplace-hldm7" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.720604 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hldm7" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.794691 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7k9df"] Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.796474 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7k9df" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.805842 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" event={"ID":"60cd23b8-6e1c-492e-aeb5-8b16609d06d1","Type":"ContainerStarted","Data":"126ca17ac1331046f38ef3e95c39b37b4252bef89c1a0511f475a2f5ad46576e"} Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.805892 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" event={"ID":"60cd23b8-6e1c-492e-aeb5-8b16609d06d1","Type":"ContainerStarted","Data":"224b2bc46f2dabb71d4e0596c8a7cb1f7d3a4099e704fb5ab5882b52edcb5f8a"} Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.806336 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.808650 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" event={"ID":"9de83f0b-7dd2-4846-a1ce-c8af930778f4","Type":"ContainerStarted","Data":"82d98e149ba56e4e52ec45de064c3a5f0fa406e38d48f08786a198d4c528ab79"} Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.810663 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" event={"ID":"ed9baf4c-70b2-450f-9b21-76dfafbc44d0","Type":"ContainerDied","Data":"1dc3c475bbed8c4530a54779f813884a49b5a06477a18a344d8eb396adfd545f"} Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.810701 4775 scope.go:117] "RemoveContainer" containerID="cadd789c9a6b6b6dccb1af86d4359d6025e42837584cd6624c1767f1cecac34c" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.810786 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.817501 4775 generic.go:334] "Generic (PLEG): container finished" podID="5bb65dee-cd5f-46b3-9e7d-36e5d182d19e" containerID="c8e5e4f89cea3a5393cd7ded536ab2f11744f74388f7645dc484cfc3075f65ca" exitCode=0 Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.817593 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-959kj" event={"ID":"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e","Type":"ContainerDied","Data":"c8e5e4f89cea3a5393cd7ded536ab2f11744f74388f7645dc484cfc3075f65ca"} Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.827099 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:16 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:16 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:16 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.827167 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.857717 4775 generic.go:334] "Generic (PLEG): container finished" podID="0f1cf896-5a0b-43e5-b36b-ba705e02825f" containerID="4b733bcb1d8da38520110c5a997c872fa32572b803088ac2576ee01c4664fef2" exitCode=0 Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.857835 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0f1cf896-5a0b-43e5-b36b-ba705e02825f","Type":"ContainerDied","Data":"4b733bcb1d8da38520110c5a997c872fa32572b803088ac2576ee01c4664fef2"} Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.862215 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" event={"ID":"f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb","Type":"ContainerDied","Data":"3aa7cdf6ad840aefd2de6cc5405b71b516de12c087774e2fad897251db2fc6ea"} Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.862257 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-r6c6z" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.864464 4775 generic.go:334] "Generic (PLEG): container finished" podID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" containerID="00e2e82d9b942bac6084d281d0485867290448469dd14201683721f46ea863fc" exitCode=0 Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.864526 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svf9w" event={"ID":"1dcbca72-150a-47c6-ac3c-f701ae82e05b","Type":"ContainerDied","Data":"00e2e82d9b942bac6084d281d0485867290448469dd14201683721f46ea863fc"} Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.870487 4775 scope.go:117] "RemoveContainer" containerID="2f124547d6c6b6e4cadb63b6c1732b81d7643f21b62f22762426569eabbfc7c4" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.871728 4775 generic.go:334] "Generic (PLEG): container finished" podID="ccb910ab-dcef-4523-81df-c0fb5eb83429" containerID="9f77f33cd00f43111cd055698fe7da46bda8e9c64a949491d07dadf580cf4426" exitCode=0 Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.871786 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbtgv" event={"ID":"ccb910ab-dcef-4523-81df-c0fb5eb83429","Type":"ContainerDied","Data":"9f77f33cd00f43111cd055698fe7da46bda8e9c64a949491d07dadf580cf4426"} Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.874684 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"829c993f-4963-416a-8d66-15cf19a68237","Type":"ContainerStarted","Data":"7a7b97953f09a11e97eecb3805a985f1f57d9595e025c3794916c5e3633729ac"} Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.874726 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"829c993f-4963-416a-8d66-15cf19a68237","Type":"ContainerStarted","Data":"8ee6fc7e5443705634d1b844abee45703b8c3f989c11bb3335bb9bbcf4253920"} Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.878942 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7k9df"] Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.893822 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct" event={"ID":"a5aa6958-e573-4efb-a031-218c62b0bec9","Type":"ContainerDied","Data":"4a3a49e3ddeff3a00ab9609e9ae3b8c22b4ec782261392142e9e4f564705d42d"} Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.893864 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a3a49e3ddeff3a00ab9609e9ae3b8c22b4ec782261392142e9e4f564705d42d" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.893938 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.903406 4775 generic.go:334] "Generic (PLEG): container finished" podID="de54a48a-b733-4042-80b6-ecc719712314" containerID="c42abfff124b04fcc4cb97eb8de832fa3a68e69db6ccb668907d78050544eb6c" exitCode=0 Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.903449 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5fgf" event={"ID":"de54a48a-b733-4042-80b6-ecc719712314","Type":"ContainerDied","Data":"c42abfff124b04fcc4cb97eb8de832fa3a68e69db6ccb668907d78050544eb6c"} Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.942478 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" podStartSLOduration=170.942457028 podStartE2EDuration="2m50.942457028s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:16.932263165 +0000 UTC m=+229.908726789" watchObservedRunningTime="2026-03-21 04:51:16.942457028 +0000 UTC m=+229.918920662" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.944664 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-hvfsp" podStartSLOduration=16.944640551 podStartE2EDuration="16.944640551s" podCreationTimestamp="2026-03-21 04:51:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:16.880144289 +0000 UTC m=+229.856607933" watchObservedRunningTime="2026-03-21 04:51:16.944640551 +0000 UTC m=+229.921104175" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.963295 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r6c6z"] Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.966293 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55e2733e-3620-4ecb-a51e-33b1fd3dce9d-catalog-content\") pod \"redhat-marketplace-7k9df\" (UID: \"55e2733e-3620-4ecb-a51e-33b1fd3dce9d\") " pod="openshift-marketplace/redhat-marketplace-7k9df" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.966329 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mm7zf\" (UniqueName: \"kubernetes.io/projected/55e2733e-3620-4ecb-a51e-33b1fd3dce9d-kube-api-access-mm7zf\") pod \"redhat-marketplace-7k9df\" (UID: \"55e2733e-3620-4ecb-a51e-33b1fd3dce9d\") " pod="openshift-marketplace/redhat-marketplace-7k9df" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.966368 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55e2733e-3620-4ecb-a51e-33b1fd3dce9d-utilities\") pod \"redhat-marketplace-7k9df\" (UID: \"55e2733e-3620-4ecb-a51e-33b1fd3dce9d\") " pod="openshift-marketplace/redhat-marketplace-7k9df" Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.971289 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r6c6z"] Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.975688 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp"] Mar 21 04:51:16 crc kubenswrapper[4775]: I0321 04:51:16.979270 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-55lcp"] Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.060957 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hldm7"] Mar 21 04:51:17 crc kubenswrapper[4775]: W0321 04:51:17.061949 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod571e84f2_a2bc_4f09_ac53_d4a4adafa80b.slice/crio-615ed550bf3a5c04a2f6727d10336a359668c9d0754381b6a62de7c552e1b1b1 WatchSource:0}: Error finding container 615ed550bf3a5c04a2f6727d10336a359668c9d0754381b6a62de7c552e1b1b1: Status 404 returned error can't find the container with id 615ed550bf3a5c04a2f6727d10336a359668c9d0754381b6a62de7c552e1b1b1 Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.063750 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.063735851 podStartE2EDuration="2.063735851s" podCreationTimestamp="2026-03-21 04:51:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:17.0563957 +0000 UTC m=+230.032859334" watchObservedRunningTime="2026-03-21 04:51:17.063735851 +0000 UTC m=+230.040199475" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.067938 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55e2733e-3620-4ecb-a51e-33b1fd3dce9d-catalog-content\") pod \"redhat-marketplace-7k9df\" (UID: \"55e2733e-3620-4ecb-a51e-33b1fd3dce9d\") " pod="openshift-marketplace/redhat-marketplace-7k9df" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.067982 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mm7zf\" (UniqueName: \"kubernetes.io/projected/55e2733e-3620-4ecb-a51e-33b1fd3dce9d-kube-api-access-mm7zf\") pod \"redhat-marketplace-7k9df\" (UID: \"55e2733e-3620-4ecb-a51e-33b1fd3dce9d\") " pod="openshift-marketplace/redhat-marketplace-7k9df" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.068031 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55e2733e-3620-4ecb-a51e-33b1fd3dce9d-utilities\") pod \"redhat-marketplace-7k9df\" (UID: \"55e2733e-3620-4ecb-a51e-33b1fd3dce9d\") " pod="openshift-marketplace/redhat-marketplace-7k9df" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.068739 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55e2733e-3620-4ecb-a51e-33b1fd3dce9d-utilities\") pod \"redhat-marketplace-7k9df\" (UID: \"55e2733e-3620-4ecb-a51e-33b1fd3dce9d\") " pod="openshift-marketplace/redhat-marketplace-7k9df" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.068762 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55e2733e-3620-4ecb-a51e-33b1fd3dce9d-catalog-content\") pod \"redhat-marketplace-7k9df\" (UID: \"55e2733e-3620-4ecb-a51e-33b1fd3dce9d\") " pod="openshift-marketplace/redhat-marketplace-7k9df" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.099441 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mm7zf\" (UniqueName: \"kubernetes.io/projected/55e2733e-3620-4ecb-a51e-33b1fd3dce9d-kube-api-access-mm7zf\") pod \"redhat-marketplace-7k9df\" (UID: \"55e2733e-3620-4ecb-a51e-33b1fd3dce9d\") " pod="openshift-marketplace/redhat-marketplace-7k9df" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.164243 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7k9df" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.191368 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wjfwd"] Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.192489 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wjfwd" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.194665 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.213778 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wjfwd"] Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.271052 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70ad413a-5f81-4094-b2d8-9b89698c6e32-utilities\") pod \"redhat-operators-wjfwd\" (UID: \"70ad413a-5f81-4094-b2d8-9b89698c6e32\") " pod="openshift-marketplace/redhat-operators-wjfwd" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.271220 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70ad413a-5f81-4094-b2d8-9b89698c6e32-catalog-content\") pod \"redhat-operators-wjfwd\" (UID: \"70ad413a-5f81-4094-b2d8-9b89698c6e32\") " pod="openshift-marketplace/redhat-operators-wjfwd" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.271356 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpppf\" (UniqueName: \"kubernetes.io/projected/70ad413a-5f81-4094-b2d8-9b89698c6e32-kube-api-access-vpppf\") pod \"redhat-operators-wjfwd\" (UID: \"70ad413a-5f81-4094-b2d8-9b89698c6e32\") " pod="openshift-marketplace/redhat-operators-wjfwd" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.372828 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70ad413a-5f81-4094-b2d8-9b89698c6e32-utilities\") pod \"redhat-operators-wjfwd\" (UID: \"70ad413a-5f81-4094-b2d8-9b89698c6e32\") " pod="openshift-marketplace/redhat-operators-wjfwd" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.373328 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70ad413a-5f81-4094-b2d8-9b89698c6e32-catalog-content\") pod \"redhat-operators-wjfwd\" (UID: \"70ad413a-5f81-4094-b2d8-9b89698c6e32\") " pod="openshift-marketplace/redhat-operators-wjfwd" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.373400 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpppf\" (UniqueName: \"kubernetes.io/projected/70ad413a-5f81-4094-b2d8-9b89698c6e32-kube-api-access-vpppf\") pod \"redhat-operators-wjfwd\" (UID: \"70ad413a-5f81-4094-b2d8-9b89698c6e32\") " pod="openshift-marketplace/redhat-operators-wjfwd" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.373715 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70ad413a-5f81-4094-b2d8-9b89698c6e32-catalog-content\") pod \"redhat-operators-wjfwd\" (UID: \"70ad413a-5f81-4094-b2d8-9b89698c6e32\") " pod="openshift-marketplace/redhat-operators-wjfwd" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.374398 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70ad413a-5f81-4094-b2d8-9b89698c6e32-utilities\") pod \"redhat-operators-wjfwd\" (UID: \"70ad413a-5f81-4094-b2d8-9b89698c6e32\") " pod="openshift-marketplace/redhat-operators-wjfwd" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.395018 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cfvcz"] Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.396046 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cfvcz" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.399781 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpppf\" (UniqueName: \"kubernetes.io/projected/70ad413a-5f81-4094-b2d8-9b89698c6e32-kube-api-access-vpppf\") pod \"redhat-operators-wjfwd\" (UID: \"70ad413a-5f81-4094-b2d8-9b89698c6e32\") " pod="openshift-marketplace/redhat-operators-wjfwd" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.402705 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cfvcz"] Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.429342 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7k9df"] Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.474567 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96069bfd-088a-4053-ab37-76a04683a6a6-catalog-content\") pod \"redhat-operators-cfvcz\" (UID: \"96069bfd-088a-4053-ab37-76a04683a6a6\") " pod="openshift-marketplace/redhat-operators-cfvcz" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.474674 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdggw\" (UniqueName: \"kubernetes.io/projected/96069bfd-088a-4053-ab37-76a04683a6a6-kube-api-access-kdggw\") pod \"redhat-operators-cfvcz\" (UID: \"96069bfd-088a-4053-ab37-76a04683a6a6\") " pod="openshift-marketplace/redhat-operators-cfvcz" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.474759 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96069bfd-088a-4053-ab37-76a04683a6a6-utilities\") pod \"redhat-operators-cfvcz\" (UID: \"96069bfd-088a-4053-ab37-76a04683a6a6\") " pod="openshift-marketplace/redhat-operators-cfvcz" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.539551 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wjfwd" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.576173 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96069bfd-088a-4053-ab37-76a04683a6a6-catalog-content\") pod \"redhat-operators-cfvcz\" (UID: \"96069bfd-088a-4053-ab37-76a04683a6a6\") " pod="openshift-marketplace/redhat-operators-cfvcz" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.576283 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdggw\" (UniqueName: \"kubernetes.io/projected/96069bfd-088a-4053-ab37-76a04683a6a6-kube-api-access-kdggw\") pod \"redhat-operators-cfvcz\" (UID: \"96069bfd-088a-4053-ab37-76a04683a6a6\") " pod="openshift-marketplace/redhat-operators-cfvcz" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.576336 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96069bfd-088a-4053-ab37-76a04683a6a6-utilities\") pod \"redhat-operators-cfvcz\" (UID: \"96069bfd-088a-4053-ab37-76a04683a6a6\") " pod="openshift-marketplace/redhat-operators-cfvcz" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.577091 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96069bfd-088a-4053-ab37-76a04683a6a6-utilities\") pod \"redhat-operators-cfvcz\" (UID: \"96069bfd-088a-4053-ab37-76a04683a6a6\") " pod="openshift-marketplace/redhat-operators-cfvcz" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.577287 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96069bfd-088a-4053-ab37-76a04683a6a6-catalog-content\") pod \"redhat-operators-cfvcz\" (UID: \"96069bfd-088a-4053-ab37-76a04683a6a6\") " pod="openshift-marketplace/redhat-operators-cfvcz" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.593404 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdggw\" (UniqueName: \"kubernetes.io/projected/96069bfd-088a-4053-ab37-76a04683a6a6-kube-api-access-kdggw\") pod \"redhat-operators-cfvcz\" (UID: \"96069bfd-088a-4053-ab37-76a04683a6a6\") " pod="openshift-marketplace/redhat-operators-cfvcz" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.669320 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.670314 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed9baf4c-70b2-450f-9b21-76dfafbc44d0" path="/var/lib/kubelet/pods/ed9baf4c-70b2-450f-9b21-76dfafbc44d0/volumes" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.670967 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb" path="/var/lib/kubelet/pods/f56fd43d-f6e7-4e8c-ab18-3d5599eefcbb/volumes" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.719397 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cfvcz" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.819102 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:17 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:17 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:17 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.819161 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.913605 4775 generic.go:334] "Generic (PLEG): container finished" podID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" containerID="ee5a58cbfa0edd60aaa49923480f7f0b63c1139c1f29a46a06572d11e904c5be" exitCode=0 Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.913674 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hldm7" event={"ID":"571e84f2-a2bc-4f09-ac53-d4a4adafa80b","Type":"ContainerDied","Data":"ee5a58cbfa0edd60aaa49923480f7f0b63c1139c1f29a46a06572d11e904c5be"} Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.913743 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hldm7" event={"ID":"571e84f2-a2bc-4f09-ac53-d4a4adafa80b","Type":"ContainerStarted","Data":"615ed550bf3a5c04a2f6727d10336a359668c9d0754381b6a62de7c552e1b1b1"} Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.916769 4775 generic.go:334] "Generic (PLEG): container finished" podID="829c993f-4963-416a-8d66-15cf19a68237" containerID="7a7b97953f09a11e97eecb3805a985f1f57d9595e025c3794916c5e3633729ac" exitCode=0 Mar 21 04:51:17 crc kubenswrapper[4775]: I0321 04:51:17.916826 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"829c993f-4963-416a-8d66-15cf19a68237","Type":"ContainerDied","Data":"7a7b97953f09a11e97eecb3805a985f1f57d9595e025c3794916c5e3633729ac"} Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.295898 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.301920 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-rz6g5" Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.560298 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-zs5pr" Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.824950 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:18 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:18 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:18 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.825039 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.870099 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-b67498df7-tct4k"] Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.871382 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.872216 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx"] Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.873016 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.877725 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.877884 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.878025 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.878044 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.878460 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.878506 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.878610 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.878648 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.879455 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.879884 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.880190 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.880319 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.886666 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.886935 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx"] Mar 21 04:51:18 crc kubenswrapper[4775]: I0321 04:51:18.891927 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-b67498df7-tct4k"] Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.004175 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcgzs\" (UniqueName: \"kubernetes.io/projected/cff31a00-04ec-4558-963b-58851e3f2a4b-kube-api-access-rcgzs\") pod \"route-controller-manager-84c957496c-2zjbx\" (UID: \"cff31a00-04ec-4558-963b-58851e3f2a4b\") " pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.004290 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1cde2211-4b17-48c3-a414-d931756a8aed-serving-cert\") pod \"controller-manager-b67498df7-tct4k\" (UID: \"1cde2211-4b17-48c3-a414-d931756a8aed\") " pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.004328 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cff31a00-04ec-4558-963b-58851e3f2a4b-serving-cert\") pod \"route-controller-manager-84c957496c-2zjbx\" (UID: \"cff31a00-04ec-4558-963b-58851e3f2a4b\") " pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.004351 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cde2211-4b17-48c3-a414-d931756a8aed-config\") pod \"controller-manager-b67498df7-tct4k\" (UID: \"1cde2211-4b17-48c3-a414-d931756a8aed\") " pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.004379 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cff31a00-04ec-4558-963b-58851e3f2a4b-client-ca\") pod \"route-controller-manager-84c957496c-2zjbx\" (UID: \"cff31a00-04ec-4558-963b-58851e3f2a4b\") " pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.004654 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff31a00-04ec-4558-963b-58851e3f2a4b-config\") pod \"route-controller-manager-84c957496c-2zjbx\" (UID: \"cff31a00-04ec-4558-963b-58851e3f2a4b\") " pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.005737 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1cde2211-4b17-48c3-a414-d931756a8aed-client-ca\") pod \"controller-manager-b67498df7-tct4k\" (UID: \"1cde2211-4b17-48c3-a414-d931756a8aed\") " pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.005794 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1cde2211-4b17-48c3-a414-d931756a8aed-proxy-ca-bundles\") pod \"controller-manager-b67498df7-tct4k\" (UID: \"1cde2211-4b17-48c3-a414-d931756a8aed\") " pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.005944 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr8fr\" (UniqueName: \"kubernetes.io/projected/1cde2211-4b17-48c3-a414-d931756a8aed-kube-api-access-pr8fr\") pod \"controller-manager-b67498df7-tct4k\" (UID: \"1cde2211-4b17-48c3-a414-d931756a8aed\") " pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.107735 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1cde2211-4b17-48c3-a414-d931756a8aed-client-ca\") pod \"controller-manager-b67498df7-tct4k\" (UID: \"1cde2211-4b17-48c3-a414-d931756a8aed\") " pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.107794 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1cde2211-4b17-48c3-a414-d931756a8aed-proxy-ca-bundles\") pod \"controller-manager-b67498df7-tct4k\" (UID: \"1cde2211-4b17-48c3-a414-d931756a8aed\") " pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.107833 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr8fr\" (UniqueName: \"kubernetes.io/projected/1cde2211-4b17-48c3-a414-d931756a8aed-kube-api-access-pr8fr\") pod \"controller-manager-b67498df7-tct4k\" (UID: \"1cde2211-4b17-48c3-a414-d931756a8aed\") " pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.107869 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcgzs\" (UniqueName: \"kubernetes.io/projected/cff31a00-04ec-4558-963b-58851e3f2a4b-kube-api-access-rcgzs\") pod \"route-controller-manager-84c957496c-2zjbx\" (UID: \"cff31a00-04ec-4558-963b-58851e3f2a4b\") " pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.107909 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1cde2211-4b17-48c3-a414-d931756a8aed-serving-cert\") pod \"controller-manager-b67498df7-tct4k\" (UID: \"1cde2211-4b17-48c3-a414-d931756a8aed\") " pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.107930 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cff31a00-04ec-4558-963b-58851e3f2a4b-serving-cert\") pod \"route-controller-manager-84c957496c-2zjbx\" (UID: \"cff31a00-04ec-4558-963b-58851e3f2a4b\") " pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.107948 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cde2211-4b17-48c3-a414-d931756a8aed-config\") pod \"controller-manager-b67498df7-tct4k\" (UID: \"1cde2211-4b17-48c3-a414-d931756a8aed\") " pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.107968 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cff31a00-04ec-4558-963b-58851e3f2a4b-client-ca\") pod \"route-controller-manager-84c957496c-2zjbx\" (UID: \"cff31a00-04ec-4558-963b-58851e3f2a4b\") " pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.108000 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff31a00-04ec-4558-963b-58851e3f2a4b-config\") pod \"route-controller-manager-84c957496c-2zjbx\" (UID: \"cff31a00-04ec-4558-963b-58851e3f2a4b\") " pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.109765 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cff31a00-04ec-4558-963b-58851e3f2a4b-client-ca\") pod \"route-controller-manager-84c957496c-2zjbx\" (UID: \"cff31a00-04ec-4558-963b-58851e3f2a4b\") " pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.110075 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff31a00-04ec-4558-963b-58851e3f2a4b-config\") pod \"route-controller-manager-84c957496c-2zjbx\" (UID: \"cff31a00-04ec-4558-963b-58851e3f2a4b\") " pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.111475 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1cde2211-4b17-48c3-a414-d931756a8aed-client-ca\") pod \"controller-manager-b67498df7-tct4k\" (UID: \"1cde2211-4b17-48c3-a414-d931756a8aed\") " pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.111545 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1cde2211-4b17-48c3-a414-d931756a8aed-proxy-ca-bundles\") pod \"controller-manager-b67498df7-tct4k\" (UID: \"1cde2211-4b17-48c3-a414-d931756a8aed\") " pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.116220 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cde2211-4b17-48c3-a414-d931756a8aed-config\") pod \"controller-manager-b67498df7-tct4k\" (UID: \"1cde2211-4b17-48c3-a414-d931756a8aed\") " pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.118162 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1cde2211-4b17-48c3-a414-d931756a8aed-serving-cert\") pod \"controller-manager-b67498df7-tct4k\" (UID: \"1cde2211-4b17-48c3-a414-d931756a8aed\") " pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.128431 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcgzs\" (UniqueName: \"kubernetes.io/projected/cff31a00-04ec-4558-963b-58851e3f2a4b-kube-api-access-rcgzs\") pod \"route-controller-manager-84c957496c-2zjbx\" (UID: \"cff31a00-04ec-4558-963b-58851e3f2a4b\") " pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.132921 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr8fr\" (UniqueName: \"kubernetes.io/projected/1cde2211-4b17-48c3-a414-d931756a8aed-kube-api-access-pr8fr\") pod \"controller-manager-b67498df7-tct4k\" (UID: \"1cde2211-4b17-48c3-a414-d931756a8aed\") " pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.137025 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cff31a00-04ec-4558-963b-58851e3f2a4b-serving-cert\") pod \"route-controller-manager-84c957496c-2zjbx\" (UID: \"cff31a00-04ec-4558-963b-58851e3f2a4b\") " pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.216482 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.226092 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.819197 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:19 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:19 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:19 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.819266 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:19 crc kubenswrapper[4775]: I0321 04:51:19.922662 4775 ???:1] "http: TLS handshake error from 192.168.126.11:34234: no serving certificate available for the kubelet" Mar 21 04:51:20 crc kubenswrapper[4775]: I0321 04:51:20.477166 4775 ???:1] "http: TLS handshake error from 192.168.126.11:34248: no serving certificate available for the kubelet" Mar 21 04:51:20 crc kubenswrapper[4775]: I0321 04:51:20.818386 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:20 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:20 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:20 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:20 crc kubenswrapper[4775]: I0321 04:51:20.818471 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:20 crc kubenswrapper[4775]: I0321 04:51:20.913623 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 21 04:51:20 crc kubenswrapper[4775]: I0321 04:51:20.939402 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"829c993f-4963-416a-8d66-15cf19a68237","Type":"ContainerDied","Data":"8ee6fc7e5443705634d1b844abee45703b8c3f989c11bb3335bb9bbcf4253920"} Mar 21 04:51:20 crc kubenswrapper[4775]: I0321 04:51:20.939450 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ee6fc7e5443705634d1b844abee45703b8c3f989c11bb3335bb9bbcf4253920" Mar 21 04:51:20 crc kubenswrapper[4775]: I0321 04:51:20.939517 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 21 04:51:21 crc kubenswrapper[4775]: I0321 04:51:21.039973 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/829c993f-4963-416a-8d66-15cf19a68237-kubelet-dir\") pod \"829c993f-4963-416a-8d66-15cf19a68237\" (UID: \"829c993f-4963-416a-8d66-15cf19a68237\") " Mar 21 04:51:21 crc kubenswrapper[4775]: I0321 04:51:21.040052 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/829c993f-4963-416a-8d66-15cf19a68237-kube-api-access\") pod \"829c993f-4963-416a-8d66-15cf19a68237\" (UID: \"829c993f-4963-416a-8d66-15cf19a68237\") " Mar 21 04:51:21 crc kubenswrapper[4775]: I0321 04:51:21.040229 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/829c993f-4963-416a-8d66-15cf19a68237-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "829c993f-4963-416a-8d66-15cf19a68237" (UID: "829c993f-4963-416a-8d66-15cf19a68237"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 04:51:21 crc kubenswrapper[4775]: I0321 04:51:21.040684 4775 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/829c993f-4963-416a-8d66-15cf19a68237-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 21 04:51:21 crc kubenswrapper[4775]: I0321 04:51:21.045293 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/829c993f-4963-416a-8d66-15cf19a68237-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "829c993f-4963-416a-8d66-15cf19a68237" (UID: "829c993f-4963-416a-8d66-15cf19a68237"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:51:21 crc kubenswrapper[4775]: I0321 04:51:21.142234 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/829c993f-4963-416a-8d66-15cf19a68237-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 04:51:21 crc kubenswrapper[4775]: I0321 04:51:21.819734 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:21 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:21 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:21 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:21 crc kubenswrapper[4775]: I0321 04:51:21.819817 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:22 crc kubenswrapper[4775]: W0321 04:51:22.276276 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55e2733e_3620_4ecb_a51e_33b1fd3dce9d.slice/crio-5477663a3a0fc934d7191139776ed4c58017ab46346a0a37602d99108fd8b86f WatchSource:0}: Error finding container 5477663a3a0fc934d7191139776ed4c58017ab46346a0a37602d99108fd8b86f: Status 404 returned error can't find the container with id 5477663a3a0fc934d7191139776ed4c58017ab46346a0a37602d99108fd8b86f Mar 21 04:51:22 crc kubenswrapper[4775]: I0321 04:51:22.320998 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 21 04:51:22 crc kubenswrapper[4775]: I0321 04:51:22.481263 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f1cf896-5a0b-43e5-b36b-ba705e02825f-kube-api-access\") pod \"0f1cf896-5a0b-43e5-b36b-ba705e02825f\" (UID: \"0f1cf896-5a0b-43e5-b36b-ba705e02825f\") " Mar 21 04:51:22 crc kubenswrapper[4775]: I0321 04:51:22.481318 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0f1cf896-5a0b-43e5-b36b-ba705e02825f-kubelet-dir\") pod \"0f1cf896-5a0b-43e5-b36b-ba705e02825f\" (UID: \"0f1cf896-5a0b-43e5-b36b-ba705e02825f\") " Mar 21 04:51:22 crc kubenswrapper[4775]: I0321 04:51:22.481798 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f1cf896-5a0b-43e5-b36b-ba705e02825f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0f1cf896-5a0b-43e5-b36b-ba705e02825f" (UID: "0f1cf896-5a0b-43e5-b36b-ba705e02825f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 04:51:22 crc kubenswrapper[4775]: I0321 04:51:22.485588 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f1cf896-5a0b-43e5-b36b-ba705e02825f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0f1cf896-5a0b-43e5-b36b-ba705e02825f" (UID: "0f1cf896-5a0b-43e5-b36b-ba705e02825f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:51:22 crc kubenswrapper[4775]: I0321 04:51:22.583523 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f1cf896-5a0b-43e5-b36b-ba705e02825f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 04:51:22 crc kubenswrapper[4775]: I0321 04:51:22.583868 4775 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0f1cf896-5a0b-43e5-b36b-ba705e02825f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 21 04:51:22 crc kubenswrapper[4775]: I0321 04:51:22.820025 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:22 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:22 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:22 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:22 crc kubenswrapper[4775]: I0321 04:51:22.820093 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:22 crc kubenswrapper[4775]: I0321 04:51:22.950926 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7k9df" event={"ID":"55e2733e-3620-4ecb-a51e-33b1fd3dce9d","Type":"ContainerStarted","Data":"5477663a3a0fc934d7191139776ed4c58017ab46346a0a37602d99108fd8b86f"} Mar 21 04:51:22 crc kubenswrapper[4775]: I0321 04:51:22.963066 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0f1cf896-5a0b-43e5-b36b-ba705e02825f","Type":"ContainerDied","Data":"9a39bc499dfd59cb84df2219ab5f515cedb58c236d866f34ae7e0567a7995091"} Mar 21 04:51:22 crc kubenswrapper[4775]: I0321 04:51:22.963103 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a39bc499dfd59cb84df2219ab5f515cedb58c236d866f34ae7e0567a7995091" Mar 21 04:51:22 crc kubenswrapper[4775]: I0321 04:51:22.963171 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 21 04:51:23 crc kubenswrapper[4775]: I0321 04:51:23.024349 4775 patch_prober.go:28] interesting pod/downloads-7954f5f757-nstnr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 21 04:51:23 crc kubenswrapper[4775]: I0321 04:51:23.024404 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nstnr" podUID="c3d58eba-4ddf-463c-baa1-1943fb60c732" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 21 04:51:23 crc kubenswrapper[4775]: I0321 04:51:23.024451 4775 patch_prober.go:28] interesting pod/downloads-7954f5f757-nstnr container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 21 04:51:23 crc kubenswrapper[4775]: I0321 04:51:23.024496 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-nstnr" podUID="c3d58eba-4ddf-463c-baa1-1943fb60c732" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 21 04:51:23 crc kubenswrapper[4775]: I0321 04:51:23.259914 4775 patch_prober.go:28] interesting pod/console-f9d7485db-wst2s container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Mar 21 04:51:23 crc kubenswrapper[4775]: I0321 04:51:23.259965 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-wst2s" podUID="a9ed1e0e-eff9-4690-bcf5-45f6074c200e" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Mar 21 04:51:23 crc kubenswrapper[4775]: I0321 04:51:23.437512 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" Mar 21 04:51:23 crc kubenswrapper[4775]: I0321 04:51:23.818769 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:23 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:23 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:23 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:23 crc kubenswrapper[4775]: I0321 04:51:23.818827 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:24 crc kubenswrapper[4775]: I0321 04:51:24.004718 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs\") pod \"network-metrics-daemon-xk9f5\" (UID: \"6920413a-2c51-466d-a16e-d14489ae0c6c\") " pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:51:24 crc kubenswrapper[4775]: I0321 04:51:24.023394 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6920413a-2c51-466d-a16e-d14489ae0c6c-metrics-certs\") pod \"network-metrics-daemon-xk9f5\" (UID: \"6920413a-2c51-466d-a16e-d14489ae0c6c\") " pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:51:24 crc kubenswrapper[4775]: I0321 04:51:24.292305 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xk9f5" Mar 21 04:51:24 crc kubenswrapper[4775]: I0321 04:51:24.359209 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wjfwd"] Mar 21 04:51:24 crc kubenswrapper[4775]: I0321 04:51:24.374967 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cfvcz"] Mar 21 04:51:24 crc kubenswrapper[4775]: I0321 04:51:24.488312 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx"] Mar 21 04:51:24 crc kubenswrapper[4775]: I0321 04:51:24.496908 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-b67498df7-tct4k"] Mar 21 04:51:24 crc kubenswrapper[4775]: W0321 04:51:24.532670 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1cde2211_4b17_48c3_a414_d931756a8aed.slice/crio-26e33d93d40acbae83163a34763d11a2400dc57b95504cce6365dbca75d924d3 WatchSource:0}: Error finding container 26e33d93d40acbae83163a34763d11a2400dc57b95504cce6365dbca75d924d3: Status 404 returned error can't find the container with id 26e33d93d40acbae83163a34763d11a2400dc57b95504cce6365dbca75d924d3 Mar 21 04:51:24 crc kubenswrapper[4775]: W0321 04:51:24.543112 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcff31a00_04ec_4558_963b_58851e3f2a4b.slice/crio-3bb4a60ce62110a0343ce33bf20b360542634587af7c8fc08808e3e39e831dc8 WatchSource:0}: Error finding container 3bb4a60ce62110a0343ce33bf20b360542634587af7c8fc08808e3e39e831dc8: Status 404 returned error can't find the container with id 3bb4a60ce62110a0343ce33bf20b360542634587af7c8fc08808e3e39e831dc8 Mar 21 04:51:24 crc kubenswrapper[4775]: I0321 04:51:24.557415 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-xk9f5"] Mar 21 04:51:24 crc kubenswrapper[4775]: W0321 04:51:24.594893 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6920413a_2c51_466d_a16e_d14489ae0c6c.slice/crio-fa19292a4f7eb2825da51194d18d10fd9a53360b9a311b276873cb1f81047c61 WatchSource:0}: Error finding container fa19292a4f7eb2825da51194d18d10fd9a53360b9a311b276873cb1f81047c61: Status 404 returned error can't find the container with id fa19292a4f7eb2825da51194d18d10fd9a53360b9a311b276873cb1f81047c61 Mar 21 04:51:24 crc kubenswrapper[4775]: I0321 04:51:24.822349 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:24 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:24 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:24 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:24 crc kubenswrapper[4775]: I0321 04:51:24.822876 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:24 crc kubenswrapper[4775]: I0321 04:51:24.984959 4775 generic.go:334] "Generic (PLEG): container finished" podID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" containerID="2fca8a116d2bdda5fb8662833d28064b756cf76d00f4ea921c85b01d826cd2be" exitCode=0 Mar 21 04:51:24 crc kubenswrapper[4775]: I0321 04:51:24.985036 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7k9df" event={"ID":"55e2733e-3620-4ecb-a51e-33b1fd3dce9d","Type":"ContainerDied","Data":"2fca8a116d2bdda5fb8662833d28064b756cf76d00f4ea921c85b01d826cd2be"} Mar 21 04:51:24 crc kubenswrapper[4775]: I0321 04:51:24.988236 4775 generic.go:334] "Generic (PLEG): container finished" podID="96069bfd-088a-4053-ab37-76a04683a6a6" containerID="e2fe2d90884acf37e8aab163e4c967e9d771b2b2c054a404750ee68dc72d022c" exitCode=0 Mar 21 04:51:24 crc kubenswrapper[4775]: I0321 04:51:24.988299 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cfvcz" event={"ID":"96069bfd-088a-4053-ab37-76a04683a6a6","Type":"ContainerDied","Data":"e2fe2d90884acf37e8aab163e4c967e9d771b2b2c054a404750ee68dc72d022c"} Mar 21 04:51:24 crc kubenswrapper[4775]: I0321 04:51:24.988325 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cfvcz" event={"ID":"96069bfd-088a-4053-ab37-76a04683a6a6","Type":"ContainerStarted","Data":"778fa4db06d864ab1638a1a9d9660eefb0a2a7f44eb1587a31bef24f86e53ae8"} Mar 21 04:51:25 crc kubenswrapper[4775]: I0321 04:51:25.014828 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" event={"ID":"6920413a-2c51-466d-a16e-d14489ae0c6c","Type":"ContainerStarted","Data":"fa19292a4f7eb2825da51194d18d10fd9a53360b9a311b276873cb1f81047c61"} Mar 21 04:51:25 crc kubenswrapper[4775]: I0321 04:51:25.025942 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" event={"ID":"1cde2211-4b17-48c3-a414-d931756a8aed","Type":"ContainerStarted","Data":"c026a9d086ef440232842548092ab223e1b082ca25729b9aaf1abf4ff41201dd"} Mar 21 04:51:25 crc kubenswrapper[4775]: I0321 04:51:25.025998 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" event={"ID":"1cde2211-4b17-48c3-a414-d931756a8aed","Type":"ContainerStarted","Data":"26e33d93d40acbae83163a34763d11a2400dc57b95504cce6365dbca75d924d3"} Mar 21 04:51:25 crc kubenswrapper[4775]: I0321 04:51:25.026236 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:51:25 crc kubenswrapper[4775]: I0321 04:51:25.032573 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567810-wb89g" event={"ID":"cd25c8a4-8047-4602-a95b-3308af65bd38","Type":"ContainerStarted","Data":"5c917ed3fbfe2c82bdf6a98ab031c5d0f484f2833ad0eb7cad0c0b48f3a9128d"} Mar 21 04:51:25 crc kubenswrapper[4775]: I0321 04:51:25.035203 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:51:25 crc kubenswrapper[4775]: I0321 04:51:25.038841 4775 generic.go:334] "Generic (PLEG): container finished" podID="70ad413a-5f81-4094-b2d8-9b89698c6e32" containerID="3b507bd384c25213b4044dd19f72806386309a77448d646faeed8c941611578b" exitCode=0 Mar 21 04:51:25 crc kubenswrapper[4775]: I0321 04:51:25.039003 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjfwd" event={"ID":"70ad413a-5f81-4094-b2d8-9b89698c6e32","Type":"ContainerDied","Data":"3b507bd384c25213b4044dd19f72806386309a77448d646faeed8c941611578b"} Mar 21 04:51:25 crc kubenswrapper[4775]: I0321 04:51:25.039098 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjfwd" event={"ID":"70ad413a-5f81-4094-b2d8-9b89698c6e32","Type":"ContainerStarted","Data":"f3ec71e4b8bce118032bd210cfae1bbf7c53e6c8a5aba3f4f045bc35e188b3c0"} Mar 21 04:51:25 crc kubenswrapper[4775]: I0321 04:51:25.045668 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" podStartSLOduration=10.045648962 podStartE2EDuration="10.045648962s" podCreationTimestamp="2026-03-21 04:51:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:25.04455906 +0000 UTC m=+238.021022694" watchObservedRunningTime="2026-03-21 04:51:25.045648962 +0000 UTC m=+238.022112586" Mar 21 04:51:25 crc kubenswrapper[4775]: I0321 04:51:25.067069 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" event={"ID":"cff31a00-04ec-4558-963b-58851e3f2a4b","Type":"ContainerStarted","Data":"adb4a7b613a2923551c3395853cd07c6da93648bbb3bbab0261ec4ce5d8387a6"} Mar 21 04:51:25 crc kubenswrapper[4775]: I0321 04:51:25.067144 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" event={"ID":"cff31a00-04ec-4558-963b-58851e3f2a4b","Type":"ContainerStarted","Data":"3bb4a60ce62110a0343ce33bf20b360542634587af7c8fc08808e3e39e831dc8"} Mar 21 04:51:25 crc kubenswrapper[4775]: I0321 04:51:25.068345 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" Mar 21 04:51:25 crc kubenswrapper[4775]: I0321 04:51:25.147828 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29567810-wb89g" podStartSLOduration=66.414388366 podStartE2EDuration="1m25.147808425s" podCreationTimestamp="2026-03-21 04:50:00 +0000 UTC" firstStartedPulling="2026-03-21 04:51:05.415107708 +0000 UTC m=+218.391571332" lastFinishedPulling="2026-03-21 04:51:24.148527767 +0000 UTC m=+237.124991391" observedRunningTime="2026-03-21 04:51:25.117833325 +0000 UTC m=+238.094296949" watchObservedRunningTime="2026-03-21 04:51:25.147808425 +0000 UTC m=+238.124272049" Mar 21 04:51:25 crc kubenswrapper[4775]: I0321 04:51:25.179728 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" podStartSLOduration=10.179707132 podStartE2EDuration="10.179707132s" podCreationTimestamp="2026-03-21 04:51:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:25.172639859 +0000 UTC m=+238.149103483" watchObservedRunningTime="2026-03-21 04:51:25.179707132 +0000 UTC m=+238.156170756" Mar 21 04:51:25 crc kubenswrapper[4775]: I0321 04:51:25.249042 4775 csr.go:261] certificate signing request csr-2dhqr is approved, waiting to be issued Mar 21 04:51:25 crc kubenswrapper[4775]: I0321 04:51:25.252828 4775 csr.go:257] certificate signing request csr-2dhqr is issued Mar 21 04:51:25 crc kubenswrapper[4775]: I0321 04:51:25.380441 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" Mar 21 04:51:25 crc kubenswrapper[4775]: I0321 04:51:25.820722 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:25 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:25 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:25 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:25 crc kubenswrapper[4775]: I0321 04:51:25.821044 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:26 crc kubenswrapper[4775]: I0321 04:51:26.085457 4775 generic.go:334] "Generic (PLEG): container finished" podID="cd25c8a4-8047-4602-a95b-3308af65bd38" containerID="5c917ed3fbfe2c82bdf6a98ab031c5d0f484f2833ad0eb7cad0c0b48f3a9128d" exitCode=0 Mar 21 04:51:26 crc kubenswrapper[4775]: I0321 04:51:26.085531 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567810-wb89g" event={"ID":"cd25c8a4-8047-4602-a95b-3308af65bd38","Type":"ContainerDied","Data":"5c917ed3fbfe2c82bdf6a98ab031c5d0f484f2833ad0eb7cad0c0b48f3a9128d"} Mar 21 04:51:26 crc kubenswrapper[4775]: I0321 04:51:26.091579 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" event={"ID":"6920413a-2c51-466d-a16e-d14489ae0c6c","Type":"ContainerStarted","Data":"3718ea3bd9f7e58a94bd2fb8fd34403fdbb7b24fef3a69e3f327b3911687ac6d"} Mar 21 04:51:26 crc kubenswrapper[4775]: I0321 04:51:26.254544 4775 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-19 20:44:19.809368839 +0000 UTC Mar 21 04:51:26 crc kubenswrapper[4775]: I0321 04:51:26.254586 4775 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6567h52m53.554785339s for next certificate rotation Mar 21 04:51:26 crc kubenswrapper[4775]: I0321 04:51:26.823343 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:26 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:26 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:26 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:26 crc kubenswrapper[4775]: I0321 04:51:26.823414 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:27 crc kubenswrapper[4775]: I0321 04:51:27.106742 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xk9f5" event={"ID":"6920413a-2c51-466d-a16e-d14489ae0c6c","Type":"ContainerStarted","Data":"f2d2a3e252ec6c2fff72b9621cb27e67c2d7171b52aeea09f4251be691845e7c"} Mar 21 04:51:27 crc kubenswrapper[4775]: I0321 04:51:27.459231 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567810-wb89g" Mar 21 04:51:27 crc kubenswrapper[4775]: I0321 04:51:27.477164 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-xk9f5" podStartSLOduration=181.477143531 podStartE2EDuration="3m1.477143531s" podCreationTimestamp="2026-03-21 04:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:51:27.122132365 +0000 UTC m=+240.098595989" watchObservedRunningTime="2026-03-21 04:51:27.477143531 +0000 UTC m=+240.453607155" Mar 21 04:51:27 crc kubenswrapper[4775]: I0321 04:51:27.592987 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gm89b\" (UniqueName: \"kubernetes.io/projected/cd25c8a4-8047-4602-a95b-3308af65bd38-kube-api-access-gm89b\") pod \"cd25c8a4-8047-4602-a95b-3308af65bd38\" (UID: \"cd25c8a4-8047-4602-a95b-3308af65bd38\") " Mar 21 04:51:27 crc kubenswrapper[4775]: I0321 04:51:27.603771 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd25c8a4-8047-4602-a95b-3308af65bd38-kube-api-access-gm89b" (OuterVolumeSpecName: "kube-api-access-gm89b") pod "cd25c8a4-8047-4602-a95b-3308af65bd38" (UID: "cd25c8a4-8047-4602-a95b-3308af65bd38"). InnerVolumeSpecName "kube-api-access-gm89b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:51:27 crc kubenswrapper[4775]: I0321 04:51:27.695986 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gm89b\" (UniqueName: \"kubernetes.io/projected/cd25c8a4-8047-4602-a95b-3308af65bd38-kube-api-access-gm89b\") on node \"crc\" DevicePath \"\"" Mar 21 04:51:27 crc kubenswrapper[4775]: I0321 04:51:27.820244 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:27 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:27 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:27 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:27 crc kubenswrapper[4775]: I0321 04:51:27.820294 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:28 crc kubenswrapper[4775]: I0321 04:51:28.117909 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567810-wb89g" event={"ID":"cd25c8a4-8047-4602-a95b-3308af65bd38","Type":"ContainerDied","Data":"05b88dca4e9352028f202490cd0cd91a11f14d92891caca0835ae23e4321056a"} Mar 21 04:51:28 crc kubenswrapper[4775]: I0321 04:51:28.117976 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05b88dca4e9352028f202490cd0cd91a11f14d92891caca0835ae23e4321056a" Mar 21 04:51:28 crc kubenswrapper[4775]: I0321 04:51:28.117932 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567810-wb89g" Mar 21 04:51:28 crc kubenswrapper[4775]: I0321 04:51:28.819602 4775 patch_prober.go:28] interesting pod/router-default-5444994796-2z46g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 04:51:28 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Mar 21 04:51:28 crc kubenswrapper[4775]: [+]process-running ok Mar 21 04:51:28 crc kubenswrapper[4775]: healthz check failed Mar 21 04:51:28 crc kubenswrapper[4775]: I0321 04:51:28.819678 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2z46g" podUID="a79bbac6-f40a-4c92-8854-7ab5e72573cc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 04:51:29 crc kubenswrapper[4775]: I0321 04:51:29.820260 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:29 crc kubenswrapper[4775]: I0321 04:51:29.823378 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-2z46g" Mar 21 04:51:32 crc kubenswrapper[4775]: I0321 04:51:32.482964 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 04:51:32 crc kubenswrapper[4775]: I0321 04:51:32.483285 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 04:51:33 crc kubenswrapper[4775]: I0321 04:51:33.038179 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-nstnr" Mar 21 04:51:33 crc kubenswrapper[4775]: I0321 04:51:33.340185 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:33 crc kubenswrapper[4775]: I0321 04:51:33.344160 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-wst2s" Mar 21 04:51:34 crc kubenswrapper[4775]: I0321 04:51:34.409775 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-b67498df7-tct4k"] Mar 21 04:51:34 crc kubenswrapper[4775]: I0321 04:51:34.409987 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" podUID="1cde2211-4b17-48c3-a414-d931756a8aed" containerName="controller-manager" containerID="cri-o://c026a9d086ef440232842548092ab223e1b082ca25729b9aaf1abf4ff41201dd" gracePeriod=30 Mar 21 04:51:34 crc kubenswrapper[4775]: I0321 04:51:34.419092 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx"] Mar 21 04:51:34 crc kubenswrapper[4775]: I0321 04:51:34.419369 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" podUID="cff31a00-04ec-4558-963b-58851e3f2a4b" containerName="route-controller-manager" containerID="cri-o://adb4a7b613a2923551c3395853cd07c6da93648bbb3bbab0261ec4ce5d8387a6" gracePeriod=30 Mar 21 04:51:35 crc kubenswrapper[4775]: I0321 04:51:35.220320 4775 generic.go:334] "Generic (PLEG): container finished" podID="1cde2211-4b17-48c3-a414-d931756a8aed" containerID="c026a9d086ef440232842548092ab223e1b082ca25729b9aaf1abf4ff41201dd" exitCode=0 Mar 21 04:51:35 crc kubenswrapper[4775]: I0321 04:51:35.220412 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" event={"ID":"1cde2211-4b17-48c3-a414-d931756a8aed","Type":"ContainerDied","Data":"c026a9d086ef440232842548092ab223e1b082ca25729b9aaf1abf4ff41201dd"} Mar 21 04:51:35 crc kubenswrapper[4775]: I0321 04:51:35.222006 4775 generic.go:334] "Generic (PLEG): container finished" podID="cff31a00-04ec-4558-963b-58851e3f2a4b" containerID="adb4a7b613a2923551c3395853cd07c6da93648bbb3bbab0261ec4ce5d8387a6" exitCode=0 Mar 21 04:51:35 crc kubenswrapper[4775]: I0321 04:51:35.222050 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" event={"ID":"cff31a00-04ec-4558-963b-58851e3f2a4b","Type":"ContainerDied","Data":"adb4a7b613a2923551c3395853cd07c6da93648bbb3bbab0261ec4ce5d8387a6"} Mar 21 04:51:36 crc kubenswrapper[4775]: I0321 04:51:36.296658 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:51:39 crc kubenswrapper[4775]: I0321 04:51:39.217031 4775 patch_prober.go:28] interesting pod/controller-manager-b67498df7-tct4k container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.55:8443/healthz\": dial tcp 10.217.0.55:8443: connect: connection refused" start-of-body= Mar 21 04:51:39 crc kubenswrapper[4775]: I0321 04:51:39.217140 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" podUID="1cde2211-4b17-48c3-a414-d931756a8aed" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.55:8443/healthz\": dial tcp 10.217.0.55:8443: connect: connection refused" Mar 21 04:51:39 crc kubenswrapper[4775]: I0321 04:51:39.226990 4775 patch_prober.go:28] interesting pod/route-controller-manager-84c957496c-2zjbx container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" start-of-body= Mar 21 04:51:39 crc kubenswrapper[4775]: I0321 04:51:39.227055 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" podUID="cff31a00-04ec-4558-963b-58851e3f2a4b" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" Mar 21 04:51:40 crc kubenswrapper[4775]: I0321 04:51:40.685520 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 04:51:43 crc kubenswrapper[4775]: I0321 04:51:43.708857 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mbgvc" Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.217361 4775 patch_prober.go:28] interesting pod/controller-manager-b67498df7-tct4k container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.55:8443/healthz\": dial tcp 10.217.0.55:8443: connect: connection refused" start-of-body= Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.219437 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" podUID="1cde2211-4b17-48c3-a414-d931756a8aed" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.55:8443/healthz\": dial tcp 10.217.0.55:8443: connect: connection refused" Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.227244 4775 patch_prober.go:28] interesting pod/route-controller-manager-84c957496c-2zjbx container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" start-of-body= Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.227306 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" podUID="cff31a00-04ec-4558-963b-58851e3f2a4b" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.370150 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 21 04:51:49 crc kubenswrapper[4775]: E0321 04:51:49.370663 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd25c8a4-8047-4602-a95b-3308af65bd38" containerName="oc" Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.370701 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd25c8a4-8047-4602-a95b-3308af65bd38" containerName="oc" Mar 21 04:51:49 crc kubenswrapper[4775]: E0321 04:51:49.370725 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f1cf896-5a0b-43e5-b36b-ba705e02825f" containerName="pruner" Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.370737 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f1cf896-5a0b-43e5-b36b-ba705e02825f" containerName="pruner" Mar 21 04:51:49 crc kubenswrapper[4775]: E0321 04:51:49.370761 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="829c993f-4963-416a-8d66-15cf19a68237" containerName="pruner" Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.370774 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="829c993f-4963-416a-8d66-15cf19a68237" containerName="pruner" Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.373304 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f1cf896-5a0b-43e5-b36b-ba705e02825f" containerName="pruner" Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.373403 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="829c993f-4963-416a-8d66-15cf19a68237" containerName="pruner" Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.373424 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd25c8a4-8047-4602-a95b-3308af65bd38" containerName="oc" Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.383150 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.383437 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.386494 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.386880 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.521255 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d6977ba5-d878-4457-a6d4-8acd42ebb089-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d6977ba5-d878-4457-a6d4-8acd42ebb089\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.521510 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d6977ba5-d878-4457-a6d4-8acd42ebb089-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d6977ba5-d878-4457-a6d4-8acd42ebb089\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.622770 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d6977ba5-d878-4457-a6d4-8acd42ebb089-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d6977ba5-d878-4457-a6d4-8acd42ebb089\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.622870 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d6977ba5-d878-4457-a6d4-8acd42ebb089-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d6977ba5-d878-4457-a6d4-8acd42ebb089\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.622943 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d6977ba5-d878-4457-a6d4-8acd42ebb089-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d6977ba5-d878-4457-a6d4-8acd42ebb089\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.647151 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d6977ba5-d878-4457-a6d4-8acd42ebb089-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d6977ba5-d878-4457-a6d4-8acd42ebb089\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 21 04:51:49 crc kubenswrapper[4775]: I0321 04:51:49.709667 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 21 04:51:54 crc kubenswrapper[4775]: I0321 04:51:54.758563 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 21 04:51:54 crc kubenswrapper[4775]: I0321 04:51:54.759839 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 21 04:51:54 crc kubenswrapper[4775]: I0321 04:51:54.772265 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 21 04:51:54 crc kubenswrapper[4775]: I0321 04:51:54.888371 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bc0e122c-1775-4bf1-9025-6288c383b3f2-var-lock\") pod \"installer-9-crc\" (UID: \"bc0e122c-1775-4bf1-9025-6288c383b3f2\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 21 04:51:54 crc kubenswrapper[4775]: I0321 04:51:54.888813 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc0e122c-1775-4bf1-9025-6288c383b3f2-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bc0e122c-1775-4bf1-9025-6288c383b3f2\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 21 04:51:54 crc kubenswrapper[4775]: I0321 04:51:54.889073 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc0e122c-1775-4bf1-9025-6288c383b3f2-kube-api-access\") pod \"installer-9-crc\" (UID: \"bc0e122c-1775-4bf1-9025-6288c383b3f2\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 21 04:51:54 crc kubenswrapper[4775]: I0321 04:51:54.990072 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc0e122c-1775-4bf1-9025-6288c383b3f2-kube-api-access\") pod \"installer-9-crc\" (UID: \"bc0e122c-1775-4bf1-9025-6288c383b3f2\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 21 04:51:54 crc kubenswrapper[4775]: I0321 04:51:54.990420 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bc0e122c-1775-4bf1-9025-6288c383b3f2-var-lock\") pod \"installer-9-crc\" (UID: \"bc0e122c-1775-4bf1-9025-6288c383b3f2\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 21 04:51:54 crc kubenswrapper[4775]: I0321 04:51:54.990528 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bc0e122c-1775-4bf1-9025-6288c383b3f2-var-lock\") pod \"installer-9-crc\" (UID: \"bc0e122c-1775-4bf1-9025-6288c383b3f2\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 21 04:51:54 crc kubenswrapper[4775]: I0321 04:51:54.990677 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc0e122c-1775-4bf1-9025-6288c383b3f2-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bc0e122c-1775-4bf1-9025-6288c383b3f2\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 21 04:51:54 crc kubenswrapper[4775]: I0321 04:51:54.990790 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc0e122c-1775-4bf1-9025-6288c383b3f2-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bc0e122c-1775-4bf1-9025-6288c383b3f2\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 21 04:51:55 crc kubenswrapper[4775]: I0321 04:51:55.009924 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc0e122c-1775-4bf1-9025-6288c383b3f2-kube-api-access\") pod \"installer-9-crc\" (UID: \"bc0e122c-1775-4bf1-9025-6288c383b3f2\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 21 04:51:55 crc kubenswrapper[4775]: I0321 04:51:55.083447 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 21 04:51:55 crc kubenswrapper[4775]: E0321 04:51:55.739817 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 21 04:51:55 crc kubenswrapper[4775]: E0321 04:51:55.740398 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-98dgw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-v5fgf_openshift-marketplace(de54a48a-b733-4042-80b6-ecc719712314): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 21 04:51:55 crc kubenswrapper[4775]: E0321 04:51:55.741562 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-v5fgf" podUID="de54a48a-b733-4042-80b6-ecc719712314" Mar 21 04:51:59 crc kubenswrapper[4775]: E0321 04:51:59.822135 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-v5fgf" podUID="de54a48a-b733-4042-80b6-ecc719712314" Mar 21 04:51:59 crc kubenswrapper[4775]: E0321 04:51:59.912005 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 21 04:51:59 crc kubenswrapper[4775]: E0321 04:51:59.912192 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kdggw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-cfvcz_openshift-marketplace(96069bfd-088a-4053-ab37-76a04683a6a6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 21 04:51:59 crc kubenswrapper[4775]: E0321 04:51:59.913392 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-cfvcz" podUID="96069bfd-088a-4053-ab37-76a04683a6a6" Mar 21 04:52:00 crc kubenswrapper[4775]: I0321 04:52:00.133109 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567812-6wppn"] Mar 21 04:52:00 crc kubenswrapper[4775]: I0321 04:52:00.135525 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567812-6wppn" Mar 21 04:52:00 crc kubenswrapper[4775]: I0321 04:52:00.137447 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 04:52:00 crc kubenswrapper[4775]: I0321 04:52:00.138239 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 04:52:00 crc kubenswrapper[4775]: I0321 04:52:00.138408 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 04:52:00 crc kubenswrapper[4775]: I0321 04:52:00.139754 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567812-6wppn"] Mar 21 04:52:00 crc kubenswrapper[4775]: I0321 04:52:00.153850 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dh2d\" (UniqueName: \"kubernetes.io/projected/d8101654-10fd-404b-ae0f-a098719418f4-kube-api-access-5dh2d\") pod \"auto-csr-approver-29567812-6wppn\" (UID: \"d8101654-10fd-404b-ae0f-a098719418f4\") " pod="openshift-infra/auto-csr-approver-29567812-6wppn" Mar 21 04:52:00 crc kubenswrapper[4775]: I0321 04:52:00.217716 4775 patch_prober.go:28] interesting pod/controller-manager-b67498df7-tct4k container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.55:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 21 04:52:00 crc kubenswrapper[4775]: I0321 04:52:00.217784 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" podUID="1cde2211-4b17-48c3-a414-d931756a8aed" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.55:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 21 04:52:00 crc kubenswrapper[4775]: I0321 04:52:00.227168 4775 patch_prober.go:28] interesting pod/route-controller-manager-84c957496c-2zjbx container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.56:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 21 04:52:00 crc kubenswrapper[4775]: I0321 04:52:00.227213 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" podUID="cff31a00-04ec-4558-963b-58851e3f2a4b" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.56:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 21 04:52:00 crc kubenswrapper[4775]: I0321 04:52:00.255601 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dh2d\" (UniqueName: \"kubernetes.io/projected/d8101654-10fd-404b-ae0f-a098719418f4-kube-api-access-5dh2d\") pod \"auto-csr-approver-29567812-6wppn\" (UID: \"d8101654-10fd-404b-ae0f-a098719418f4\") " pod="openshift-infra/auto-csr-approver-29567812-6wppn" Mar 21 04:52:00 crc kubenswrapper[4775]: I0321 04:52:00.279396 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dh2d\" (UniqueName: \"kubernetes.io/projected/d8101654-10fd-404b-ae0f-a098719418f4-kube-api-access-5dh2d\") pod \"auto-csr-approver-29567812-6wppn\" (UID: \"d8101654-10fd-404b-ae0f-a098719418f4\") " pod="openshift-infra/auto-csr-approver-29567812-6wppn" Mar 21 04:52:00 crc kubenswrapper[4775]: I0321 04:52:00.458197 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567812-6wppn" Mar 21 04:52:01 crc kubenswrapper[4775]: E0321 04:52:01.485051 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-cfvcz" podUID="96069bfd-088a-4053-ab37-76a04683a6a6" Mar 21 04:52:01 crc kubenswrapper[4775]: E0321 04:52:01.513400 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 21 04:52:01 crc kubenswrapper[4775]: E0321 04:52:01.513587 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vpppf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-wjfwd_openshift-marketplace(70ad413a-5f81-4094-b2d8-9b89698c6e32): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 21 04:52:01 crc kubenswrapper[4775]: E0321 04:52:01.514840 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-wjfwd" podUID="70ad413a-5f81-4094-b2d8-9b89698c6e32" Mar 21 04:52:01 crc kubenswrapper[4775]: E0321 04:52:01.566378 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 21 04:52:01 crc kubenswrapper[4775]: E0321 04:52:01.566682 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-92fzw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-wbtgv_openshift-marketplace(ccb910ab-dcef-4523-81df-c0fb5eb83429): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 21 04:52:01 crc kubenswrapper[4775]: E0321 04:52:01.568413 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-wbtgv" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" Mar 21 04:52:01 crc kubenswrapper[4775]: E0321 04:52:01.579462 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 21 04:52:01 crc kubenswrapper[4775]: E0321 04:52:01.579646 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7ww6h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-959kj_openshift-marketplace(5bb65dee-cd5f-46b3-9e7d-36e5d182d19e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 21 04:52:01 crc kubenswrapper[4775]: E0321 04:52:01.580828 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-959kj" podUID="5bb65dee-cd5f-46b3-9e7d-36e5d182d19e" Mar 21 04:52:02 crc kubenswrapper[4775]: I0321 04:52:02.482040 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 04:52:02 crc kubenswrapper[4775]: I0321 04:52:02.482455 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 04:52:02 crc kubenswrapper[4775]: I0321 04:52:02.482513 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 04:52:02 crc kubenswrapper[4775]: I0321 04:52:02.483167 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee"} pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 04:52:02 crc kubenswrapper[4775]: I0321 04:52:02.483234 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" containerID="cri-o://94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee" gracePeriod=600 Mar 21 04:52:03 crc kubenswrapper[4775]: E0321 04:52:03.144504 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-wjfwd" podUID="70ad413a-5f81-4094-b2d8-9b89698c6e32" Mar 21 04:52:03 crc kubenswrapper[4775]: E0321 04:52:03.144504 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-wbtgv" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" Mar 21 04:52:03 crc kubenswrapper[4775]: E0321 04:52:03.144532 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-959kj" podUID="5bb65dee-cd5f-46b3-9e7d-36e5d182d19e" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.293841 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.302317 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cff31a00-04ec-4558-963b-58851e3f2a4b-client-ca\") pod \"cff31a00-04ec-4558-963b-58851e3f2a4b\" (UID: \"cff31a00-04ec-4558-963b-58851e3f2a4b\") " Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.302389 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cff31a00-04ec-4558-963b-58851e3f2a4b-serving-cert\") pod \"cff31a00-04ec-4558-963b-58851e3f2a4b\" (UID: \"cff31a00-04ec-4558-963b-58851e3f2a4b\") " Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.302429 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcgzs\" (UniqueName: \"kubernetes.io/projected/cff31a00-04ec-4558-963b-58851e3f2a4b-kube-api-access-rcgzs\") pod \"cff31a00-04ec-4558-963b-58851e3f2a4b\" (UID: \"cff31a00-04ec-4558-963b-58851e3f2a4b\") " Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.302473 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff31a00-04ec-4558-963b-58851e3f2a4b-config\") pod \"cff31a00-04ec-4558-963b-58851e3f2a4b\" (UID: \"cff31a00-04ec-4558-963b-58851e3f2a4b\") " Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.304027 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cff31a00-04ec-4558-963b-58851e3f2a4b-client-ca" (OuterVolumeSpecName: "client-ca") pod "cff31a00-04ec-4558-963b-58851e3f2a4b" (UID: "cff31a00-04ec-4558-963b-58851e3f2a4b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.305779 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cff31a00-04ec-4558-963b-58851e3f2a4b-config" (OuterVolumeSpecName: "config") pod "cff31a00-04ec-4558-963b-58851e3f2a4b" (UID: "cff31a00-04ec-4558-963b-58851e3f2a4b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.314900 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cff31a00-04ec-4558-963b-58851e3f2a4b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "cff31a00-04ec-4558-963b-58851e3f2a4b" (UID: "cff31a00-04ec-4558-963b-58851e3f2a4b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.332637 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.334225 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cff31a00-04ec-4558-963b-58851e3f2a4b-kube-api-access-rcgzs" (OuterVolumeSpecName: "kube-api-access-rcgzs") pod "cff31a00-04ec-4558-963b-58851e3f2a4b" (UID: "cff31a00-04ec-4558-963b-58851e3f2a4b"). InnerVolumeSpecName "kube-api-access-rcgzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.336515 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8"] Mar 21 04:52:03 crc kubenswrapper[4775]: E0321 04:52:03.336834 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cde2211-4b17-48c3-a414-d931756a8aed" containerName="controller-manager" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.336851 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cde2211-4b17-48c3-a414-d931756a8aed" containerName="controller-manager" Mar 21 04:52:03 crc kubenswrapper[4775]: E0321 04:52:03.336866 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cff31a00-04ec-4558-963b-58851e3f2a4b" containerName="route-controller-manager" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.336872 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="cff31a00-04ec-4558-963b-58851e3f2a4b" containerName="route-controller-manager" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.336980 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="cff31a00-04ec-4558-963b-58851e3f2a4b" containerName="route-controller-manager" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.336995 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cde2211-4b17-48c3-a414-d931756a8aed" containerName="controller-manager" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.337682 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.348442 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8"] Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.393928 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" event={"ID":"1cde2211-4b17-48c3-a414-d931756a8aed","Type":"ContainerDied","Data":"26e33d93d40acbae83163a34763d11a2400dc57b95504cce6365dbca75d924d3"} Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.393976 4775 scope.go:117] "RemoveContainer" containerID="c026a9d086ef440232842548092ab223e1b082ca25729b9aaf1abf4ff41201dd" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.394072 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b67498df7-tct4k" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.399982 4775 generic.go:334] "Generic (PLEG): container finished" podID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerID="94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee" exitCode=0 Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.400048 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerDied","Data":"94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee"} Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.402968 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1cde2211-4b17-48c3-a414-d931756a8aed-serving-cert\") pod \"1cde2211-4b17-48c3-a414-d931756a8aed\" (UID: \"1cde2211-4b17-48c3-a414-d931756a8aed\") " Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.403248 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1cde2211-4b17-48c3-a414-d931756a8aed-proxy-ca-bundles\") pod \"1cde2211-4b17-48c3-a414-d931756a8aed\" (UID: \"1cde2211-4b17-48c3-a414-d931756a8aed\") " Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.403290 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1cde2211-4b17-48c3-a414-d931756a8aed-client-ca\") pod \"1cde2211-4b17-48c3-a414-d931756a8aed\" (UID: \"1cde2211-4b17-48c3-a414-d931756a8aed\") " Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.403313 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pr8fr\" (UniqueName: \"kubernetes.io/projected/1cde2211-4b17-48c3-a414-d931756a8aed-kube-api-access-pr8fr\") pod \"1cde2211-4b17-48c3-a414-d931756a8aed\" (UID: \"1cde2211-4b17-48c3-a414-d931756a8aed\") " Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.403372 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cde2211-4b17-48c3-a414-d931756a8aed-config\") pod \"1cde2211-4b17-48c3-a414-d931756a8aed\" (UID: \"1cde2211-4b17-48c3-a414-d931756a8aed\") " Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.403494 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ef0e747e-48cd-4a7d-922d-905e15ea750e-client-ca\") pod \"route-controller-manager-7fbcc6866-hpvz8\" (UID: \"ef0e747e-48cd-4a7d-922d-905e15ea750e\") " pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.403530 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hc9l\" (UniqueName: \"kubernetes.io/projected/ef0e747e-48cd-4a7d-922d-905e15ea750e-kube-api-access-9hc9l\") pod \"route-controller-manager-7fbcc6866-hpvz8\" (UID: \"ef0e747e-48cd-4a7d-922d-905e15ea750e\") " pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.403552 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef0e747e-48cd-4a7d-922d-905e15ea750e-serving-cert\") pod \"route-controller-manager-7fbcc6866-hpvz8\" (UID: \"ef0e747e-48cd-4a7d-922d-905e15ea750e\") " pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.403611 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef0e747e-48cd-4a7d-922d-905e15ea750e-config\") pod \"route-controller-manager-7fbcc6866-hpvz8\" (UID: \"ef0e747e-48cd-4a7d-922d-905e15ea750e\") " pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.403685 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cff31a00-04ec-4558-963b-58851e3f2a4b-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.403698 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cff31a00-04ec-4558-963b-58851e3f2a4b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.403707 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcgzs\" (UniqueName: \"kubernetes.io/projected/cff31a00-04ec-4558-963b-58851e3f2a4b-kube-api-access-rcgzs\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.403716 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff31a00-04ec-4558-963b-58851e3f2a4b-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.404252 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cde2211-4b17-48c3-a414-d931756a8aed-client-ca" (OuterVolumeSpecName: "client-ca") pod "1cde2211-4b17-48c3-a414-d931756a8aed" (UID: "1cde2211-4b17-48c3-a414-d931756a8aed"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.404500 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cde2211-4b17-48c3-a414-d931756a8aed-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1cde2211-4b17-48c3-a414-d931756a8aed" (UID: "1cde2211-4b17-48c3-a414-d931756a8aed"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.405232 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cde2211-4b17-48c3-a414-d931756a8aed-config" (OuterVolumeSpecName: "config") pod "1cde2211-4b17-48c3-a414-d931756a8aed" (UID: "1cde2211-4b17-48c3-a414-d931756a8aed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.405953 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" event={"ID":"cff31a00-04ec-4558-963b-58851e3f2a4b","Type":"ContainerDied","Data":"3bb4a60ce62110a0343ce33bf20b360542634587af7c8fc08808e3e39e831dc8"} Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.406047 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.407599 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cde2211-4b17-48c3-a414-d931756a8aed-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1cde2211-4b17-48c3-a414-d931756a8aed" (UID: "1cde2211-4b17-48c3-a414-d931756a8aed"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.408903 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cde2211-4b17-48c3-a414-d931756a8aed-kube-api-access-pr8fr" (OuterVolumeSpecName: "kube-api-access-pr8fr") pod "1cde2211-4b17-48c3-a414-d931756a8aed" (UID: "1cde2211-4b17-48c3-a414-d931756a8aed"). InnerVolumeSpecName "kube-api-access-pr8fr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.425342 4775 scope.go:117] "RemoveContainer" containerID="adb4a7b613a2923551c3395853cd07c6da93648bbb3bbab0261ec4ce5d8387a6" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.435146 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx"] Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.438542 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84c957496c-2zjbx"] Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.504585 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ef0e747e-48cd-4a7d-922d-905e15ea750e-client-ca\") pod \"route-controller-manager-7fbcc6866-hpvz8\" (UID: \"ef0e747e-48cd-4a7d-922d-905e15ea750e\") " pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.504630 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hc9l\" (UniqueName: \"kubernetes.io/projected/ef0e747e-48cd-4a7d-922d-905e15ea750e-kube-api-access-9hc9l\") pod \"route-controller-manager-7fbcc6866-hpvz8\" (UID: \"ef0e747e-48cd-4a7d-922d-905e15ea750e\") " pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.504656 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef0e747e-48cd-4a7d-922d-905e15ea750e-serving-cert\") pod \"route-controller-manager-7fbcc6866-hpvz8\" (UID: \"ef0e747e-48cd-4a7d-922d-905e15ea750e\") " pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.504690 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef0e747e-48cd-4a7d-922d-905e15ea750e-config\") pod \"route-controller-manager-7fbcc6866-hpvz8\" (UID: \"ef0e747e-48cd-4a7d-922d-905e15ea750e\") " pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.504783 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cde2211-4b17-48c3-a414-d931756a8aed-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.504794 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1cde2211-4b17-48c3-a414-d931756a8aed-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.504806 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1cde2211-4b17-48c3-a414-d931756a8aed-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.504815 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1cde2211-4b17-48c3-a414-d931756a8aed-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.504826 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pr8fr\" (UniqueName: \"kubernetes.io/projected/1cde2211-4b17-48c3-a414-d931756a8aed-kube-api-access-pr8fr\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.506204 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef0e747e-48cd-4a7d-922d-905e15ea750e-config\") pod \"route-controller-manager-7fbcc6866-hpvz8\" (UID: \"ef0e747e-48cd-4a7d-922d-905e15ea750e\") " pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.506507 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ef0e747e-48cd-4a7d-922d-905e15ea750e-client-ca\") pod \"route-controller-manager-7fbcc6866-hpvz8\" (UID: \"ef0e747e-48cd-4a7d-922d-905e15ea750e\") " pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.508588 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef0e747e-48cd-4a7d-922d-905e15ea750e-serving-cert\") pod \"route-controller-manager-7fbcc6866-hpvz8\" (UID: \"ef0e747e-48cd-4a7d-922d-905e15ea750e\") " pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.525853 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hc9l\" (UniqueName: \"kubernetes.io/projected/ef0e747e-48cd-4a7d-922d-905e15ea750e-kube-api-access-9hc9l\") pod \"route-controller-manager-7fbcc6866-hpvz8\" (UID: \"ef0e747e-48cd-4a7d-922d-905e15ea750e\") " pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.645502 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.655901 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 21 04:52:03 crc kubenswrapper[4775]: W0321 04:52:03.658711 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podbc0e122c_1775_4bf1_9025_6288c383b3f2.slice/crio-df0920ca1c2002e3d0e08d9fcffa648209bafd25e83252350dced0f61be9bf09 WatchSource:0}: Error finding container df0920ca1c2002e3d0e08d9fcffa648209bafd25e83252350dced0f61be9bf09: Status 404 returned error can't find the container with id df0920ca1c2002e3d0e08d9fcffa648209bafd25e83252350dced0f61be9bf09 Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.667568 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cff31a00-04ec-4558-963b-58851e3f2a4b" path="/var/lib/kubelet/pods/cff31a00-04ec-4558-963b-58851e3f2a4b/volumes" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.670335 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.711207 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-b67498df7-tct4k"] Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.714787 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-b67498df7-tct4k"] Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.724906 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567812-6wppn"] Mar 21 04:52:03 crc kubenswrapper[4775]: W0321 04:52:03.736943 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8101654_10fd_404b_ae0f_a098719418f4.slice/crio-b4836b0688f8984a0812fbf1a29d1435aec6aa3f8171f16c027c71eb201157a9 WatchSource:0}: Error finding container b4836b0688f8984a0812fbf1a29d1435aec6aa3f8171f16c027c71eb201157a9: Status 404 returned error can't find the container with id b4836b0688f8984a0812fbf1a29d1435aec6aa3f8171f16c027c71eb201157a9 Mar 21 04:52:03 crc kubenswrapper[4775]: I0321 04:52:03.899152 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8"] Mar 21 04:52:04 crc kubenswrapper[4775]: E0321 04:52:04.373332 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 21 04:52:04 crc kubenswrapper[4775]: E0321 04:52:04.373761 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mm7zf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-7k9df_openshift-marketplace(55e2733e-3620-4ecb-a51e-33b1fd3dce9d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 21 04:52:04 crc kubenswrapper[4775]: E0321 04:52:04.374652 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 21 04:52:04 crc kubenswrapper[4775]: E0321 04:52:04.374941 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2gdv6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-hldm7_openshift-marketplace(571e84f2-a2bc-4f09-ac53-d4a4adafa80b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 21 04:52:04 crc kubenswrapper[4775]: E0321 04:52:04.374953 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-7k9df" podUID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" Mar 21 04:52:04 crc kubenswrapper[4775]: E0321 04:52:04.377475 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-hldm7" podUID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" Mar 21 04:52:04 crc kubenswrapper[4775]: I0321 04:52:04.412840 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bc0e122c-1775-4bf1-9025-6288c383b3f2","Type":"ContainerStarted","Data":"de0bb723a8e5a2f1ecac5f9d70f09000ecc4cbf0b2ce82838fe0f1932e47f097"} Mar 21 04:52:04 crc kubenswrapper[4775]: I0321 04:52:04.412888 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bc0e122c-1775-4bf1-9025-6288c383b3f2","Type":"ContainerStarted","Data":"df0920ca1c2002e3d0e08d9fcffa648209bafd25e83252350dced0f61be9bf09"} Mar 21 04:52:04 crc kubenswrapper[4775]: I0321 04:52:04.414921 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"d6977ba5-d878-4457-a6d4-8acd42ebb089","Type":"ContainerStarted","Data":"49f03f37f3492994f82774da13ca12406fd7976bb8941707a7ec3e837116fcba"} Mar 21 04:52:04 crc kubenswrapper[4775]: I0321 04:52:04.414957 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"d6977ba5-d878-4457-a6d4-8acd42ebb089","Type":"ContainerStarted","Data":"bb2095279e346ad9805b3d902d26b2c1ec43b438031bbbe2a123459d83839e23"} Mar 21 04:52:04 crc kubenswrapper[4775]: I0321 04:52:04.416628 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567812-6wppn" event={"ID":"d8101654-10fd-404b-ae0f-a098719418f4","Type":"ContainerStarted","Data":"b4836b0688f8984a0812fbf1a29d1435aec6aa3f8171f16c027c71eb201157a9"} Mar 21 04:52:04 crc kubenswrapper[4775]: I0321 04:52:04.418560 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerStarted","Data":"bce245e457399f7eaba83f68d2114b3ba9e57f5a921fcddf334a4766f16c7398"} Mar 21 04:52:04 crc kubenswrapper[4775]: I0321 04:52:04.420489 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" event={"ID":"ef0e747e-48cd-4a7d-922d-905e15ea750e","Type":"ContainerStarted","Data":"2dfd1a63e89a57a5a6969d7c49601a2ff07f69ec16bc071f9034c83d8f03b10b"} Mar 21 04:52:04 crc kubenswrapper[4775]: I0321 04:52:04.420530 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" event={"ID":"ef0e747e-48cd-4a7d-922d-905e15ea750e","Type":"ContainerStarted","Data":"f1c5e9ed26c65634fec4b7e068160478cde4bffe6efaa83fc8c27834b2f650f2"} Mar 21 04:52:04 crc kubenswrapper[4775]: E0321 04:52:04.422470 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-hldm7" podUID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" Mar 21 04:52:04 crc kubenswrapper[4775]: E0321 04:52:04.422471 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7k9df" podUID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" Mar 21 04:52:04 crc kubenswrapper[4775]: I0321 04:52:04.432978 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=10.432957254 podStartE2EDuration="10.432957254s" podCreationTimestamp="2026-03-21 04:51:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:52:04.430425181 +0000 UTC m=+277.406888825" watchObservedRunningTime="2026-03-21 04:52:04.432957254 +0000 UTC m=+277.409420888" Mar 21 04:52:04 crc kubenswrapper[4775]: I0321 04:52:04.463134 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=15.463101735 podStartE2EDuration="15.463101735s" podCreationTimestamp="2026-03-21 04:51:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:52:04.461307743 +0000 UTC m=+277.437771377" watchObservedRunningTime="2026-03-21 04:52:04.463101735 +0000 UTC m=+277.439565359" Mar 21 04:52:04 crc kubenswrapper[4775]: E0321 04:52:04.480789 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 21 04:52:04 crc kubenswrapper[4775]: E0321 04:52:04.480924 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4wq48,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-svf9w_openshift-marketplace(1dcbca72-150a-47c6-ac3c-f701ae82e05b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 21 04:52:04 crc kubenswrapper[4775]: E0321 04:52:04.482273 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-svf9w" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" Mar 21 04:52:04 crc kubenswrapper[4775]: I0321 04:52:04.483444 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" podStartSLOduration=10.483431752 podStartE2EDuration="10.483431752s" podCreationTimestamp="2026-03-21 04:51:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:52:04.480318702 +0000 UTC m=+277.456782326" watchObservedRunningTime="2026-03-21 04:52:04.483431752 +0000 UTC m=+277.459895386" Mar 21 04:52:05 crc kubenswrapper[4775]: I0321 04:52:05.427992 4775 generic.go:334] "Generic (PLEG): container finished" podID="d8101654-10fd-404b-ae0f-a098719418f4" containerID="4cd8fe98e0e79605364ea6850bbec4e960085c0c48458cb67fb164296fe16045" exitCode=0 Mar 21 04:52:05 crc kubenswrapper[4775]: I0321 04:52:05.428033 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567812-6wppn" event={"ID":"d8101654-10fd-404b-ae0f-a098719418f4","Type":"ContainerDied","Data":"4cd8fe98e0e79605364ea6850bbec4e960085c0c48458cb67fb164296fe16045"} Mar 21 04:52:05 crc kubenswrapper[4775]: I0321 04:52:05.429760 4775 generic.go:334] "Generic (PLEG): container finished" podID="d6977ba5-d878-4457-a6d4-8acd42ebb089" containerID="49f03f37f3492994f82774da13ca12406fd7976bb8941707a7ec3e837116fcba" exitCode=0 Mar 21 04:52:05 crc kubenswrapper[4775]: I0321 04:52:05.429890 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"d6977ba5-d878-4457-a6d4-8acd42ebb089","Type":"ContainerDied","Data":"49f03f37f3492994f82774da13ca12406fd7976bb8941707a7ec3e837116fcba"} Mar 21 04:52:05 crc kubenswrapper[4775]: I0321 04:52:05.430078 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" Mar 21 04:52:05 crc kubenswrapper[4775]: E0321 04:52:05.432717 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-svf9w" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" Mar 21 04:52:05 crc kubenswrapper[4775]: I0321 04:52:05.437243 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" Mar 21 04:52:05 crc kubenswrapper[4775]: I0321 04:52:05.667864 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cde2211-4b17-48c3-a414-d931756a8aed" path="/var/lib/kubelet/pods/1cde2211-4b17-48c3-a414-d931756a8aed/volumes" Mar 21 04:52:05 crc kubenswrapper[4775]: I0321 04:52:05.946206 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7bd4667b97-z2cn5"] Mar 21 04:52:05 crc kubenswrapper[4775]: I0321 04:52:05.946874 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:05 crc kubenswrapper[4775]: I0321 04:52:05.948739 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 21 04:52:05 crc kubenswrapper[4775]: I0321 04:52:05.949026 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 21 04:52:05 crc kubenswrapper[4775]: I0321 04:52:05.952380 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 21 04:52:05 crc kubenswrapper[4775]: I0321 04:52:05.952616 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 21 04:52:05 crc kubenswrapper[4775]: I0321 04:52:05.953237 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 21 04:52:05 crc kubenswrapper[4775]: I0321 04:52:05.953494 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 21 04:52:05 crc kubenswrapper[4775]: I0321 04:52:05.957812 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 21 04:52:05 crc kubenswrapper[4775]: I0321 04:52:05.958882 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7bd4667b97-z2cn5"] Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.139306 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6s67\" (UniqueName: \"kubernetes.io/projected/05791985-4754-4e1e-9400-7204c6a5eab2-kube-api-access-l6s67\") pod \"controller-manager-7bd4667b97-z2cn5\" (UID: \"05791985-4754-4e1e-9400-7204c6a5eab2\") " pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.139352 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05791985-4754-4e1e-9400-7204c6a5eab2-config\") pod \"controller-manager-7bd4667b97-z2cn5\" (UID: \"05791985-4754-4e1e-9400-7204c6a5eab2\") " pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.139384 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05791985-4754-4e1e-9400-7204c6a5eab2-serving-cert\") pod \"controller-manager-7bd4667b97-z2cn5\" (UID: \"05791985-4754-4e1e-9400-7204c6a5eab2\") " pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.139412 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05791985-4754-4e1e-9400-7204c6a5eab2-client-ca\") pod \"controller-manager-7bd4667b97-z2cn5\" (UID: \"05791985-4754-4e1e-9400-7204c6a5eab2\") " pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.139444 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/05791985-4754-4e1e-9400-7204c6a5eab2-proxy-ca-bundles\") pod \"controller-manager-7bd4667b97-z2cn5\" (UID: \"05791985-4754-4e1e-9400-7204c6a5eab2\") " pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.254793 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6s67\" (UniqueName: \"kubernetes.io/projected/05791985-4754-4e1e-9400-7204c6a5eab2-kube-api-access-l6s67\") pod \"controller-manager-7bd4667b97-z2cn5\" (UID: \"05791985-4754-4e1e-9400-7204c6a5eab2\") " pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.254875 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05791985-4754-4e1e-9400-7204c6a5eab2-config\") pod \"controller-manager-7bd4667b97-z2cn5\" (UID: \"05791985-4754-4e1e-9400-7204c6a5eab2\") " pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.254947 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05791985-4754-4e1e-9400-7204c6a5eab2-serving-cert\") pod \"controller-manager-7bd4667b97-z2cn5\" (UID: \"05791985-4754-4e1e-9400-7204c6a5eab2\") " pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.255022 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05791985-4754-4e1e-9400-7204c6a5eab2-client-ca\") pod \"controller-manager-7bd4667b97-z2cn5\" (UID: \"05791985-4754-4e1e-9400-7204c6a5eab2\") " pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.255168 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/05791985-4754-4e1e-9400-7204c6a5eab2-proxy-ca-bundles\") pod \"controller-manager-7bd4667b97-z2cn5\" (UID: \"05791985-4754-4e1e-9400-7204c6a5eab2\") " pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.256837 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05791985-4754-4e1e-9400-7204c6a5eab2-client-ca\") pod \"controller-manager-7bd4667b97-z2cn5\" (UID: \"05791985-4754-4e1e-9400-7204c6a5eab2\") " pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.256896 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/05791985-4754-4e1e-9400-7204c6a5eab2-proxy-ca-bundles\") pod \"controller-manager-7bd4667b97-z2cn5\" (UID: \"05791985-4754-4e1e-9400-7204c6a5eab2\") " pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.257213 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05791985-4754-4e1e-9400-7204c6a5eab2-config\") pod \"controller-manager-7bd4667b97-z2cn5\" (UID: \"05791985-4754-4e1e-9400-7204c6a5eab2\") " pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.262649 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05791985-4754-4e1e-9400-7204c6a5eab2-serving-cert\") pod \"controller-manager-7bd4667b97-z2cn5\" (UID: \"05791985-4754-4e1e-9400-7204c6a5eab2\") " pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.275061 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6s67\" (UniqueName: \"kubernetes.io/projected/05791985-4754-4e1e-9400-7204c6a5eab2-kube-api-access-l6s67\") pod \"controller-manager-7bd4667b97-z2cn5\" (UID: \"05791985-4754-4e1e-9400-7204c6a5eab2\") " pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.572446 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.595483 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jd86b"] Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.741907 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.764048 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d6977ba5-d878-4457-a6d4-8acd42ebb089-kubelet-dir\") pod \"d6977ba5-d878-4457-a6d4-8acd42ebb089\" (UID: \"d6977ba5-d878-4457-a6d4-8acd42ebb089\") " Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.764127 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d6977ba5-d878-4457-a6d4-8acd42ebb089-kube-api-access\") pod \"d6977ba5-d878-4457-a6d4-8acd42ebb089\" (UID: \"d6977ba5-d878-4457-a6d4-8acd42ebb089\") " Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.766348 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6977ba5-d878-4457-a6d4-8acd42ebb089-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d6977ba5-d878-4457-a6d4-8acd42ebb089" (UID: "d6977ba5-d878-4457-a6d4-8acd42ebb089"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.776281 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6977ba5-d878-4457-a6d4-8acd42ebb089-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d6977ba5-d878-4457-a6d4-8acd42ebb089" (UID: "d6977ba5-d878-4457-a6d4-8acd42ebb089"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.849013 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567812-6wppn" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.865822 4775 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d6977ba5-d878-4457-a6d4-8acd42ebb089-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.865857 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d6977ba5-d878-4457-a6d4-8acd42ebb089-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.967184 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dh2d\" (UniqueName: \"kubernetes.io/projected/d8101654-10fd-404b-ae0f-a098719418f4-kube-api-access-5dh2d\") pod \"d8101654-10fd-404b-ae0f-a098719418f4\" (UID: \"d8101654-10fd-404b-ae0f-a098719418f4\") " Mar 21 04:52:06 crc kubenswrapper[4775]: I0321 04:52:06.970620 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8101654-10fd-404b-ae0f-a098719418f4-kube-api-access-5dh2d" (OuterVolumeSpecName: "kube-api-access-5dh2d") pod "d8101654-10fd-404b-ae0f-a098719418f4" (UID: "d8101654-10fd-404b-ae0f-a098719418f4"). InnerVolumeSpecName "kube-api-access-5dh2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:52:07 crc kubenswrapper[4775]: I0321 04:52:07.068489 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dh2d\" (UniqueName: \"kubernetes.io/projected/d8101654-10fd-404b-ae0f-a098719418f4-kube-api-access-5dh2d\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:07 crc kubenswrapper[4775]: I0321 04:52:07.162927 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7bd4667b97-z2cn5"] Mar 21 04:52:07 crc kubenswrapper[4775]: W0321 04:52:07.173478 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05791985_4754_4e1e_9400_7204c6a5eab2.slice/crio-1b352205937c3361b5c47711682f5ba564f7c8d78431497ed52ef09e4be45784 WatchSource:0}: Error finding container 1b352205937c3361b5c47711682f5ba564f7c8d78431497ed52ef09e4be45784: Status 404 returned error can't find the container with id 1b352205937c3361b5c47711682f5ba564f7c8d78431497ed52ef09e4be45784 Mar 21 04:52:07 crc kubenswrapper[4775]: I0321 04:52:07.441637 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"d6977ba5-d878-4457-a6d4-8acd42ebb089","Type":"ContainerDied","Data":"bb2095279e346ad9805b3d902d26b2c1ec43b438031bbbe2a123459d83839e23"} Mar 21 04:52:07 crc kubenswrapper[4775]: I0321 04:52:07.441681 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb2095279e346ad9805b3d902d26b2c1ec43b438031bbbe2a123459d83839e23" Mar 21 04:52:07 crc kubenswrapper[4775]: I0321 04:52:07.441742 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 21 04:52:07 crc kubenswrapper[4775]: I0321 04:52:07.446505 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567812-6wppn" Mar 21 04:52:07 crc kubenswrapper[4775]: I0321 04:52:07.446504 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567812-6wppn" event={"ID":"d8101654-10fd-404b-ae0f-a098719418f4","Type":"ContainerDied","Data":"b4836b0688f8984a0812fbf1a29d1435aec6aa3f8171f16c027c71eb201157a9"} Mar 21 04:52:07 crc kubenswrapper[4775]: I0321 04:52:07.446639 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4836b0688f8984a0812fbf1a29d1435aec6aa3f8171f16c027c71eb201157a9" Mar 21 04:52:07 crc kubenswrapper[4775]: I0321 04:52:07.449359 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" event={"ID":"05791985-4754-4e1e-9400-7204c6a5eab2","Type":"ContainerStarted","Data":"5494d63ffe4e008a81f5adf939d712754549a2efc64457da115e33ba5aea120c"} Mar 21 04:52:07 crc kubenswrapper[4775]: I0321 04:52:07.449401 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" event={"ID":"05791985-4754-4e1e-9400-7204c6a5eab2","Type":"ContainerStarted","Data":"1b352205937c3361b5c47711682f5ba564f7c8d78431497ed52ef09e4be45784"} Mar 21 04:52:07 crc kubenswrapper[4775]: I0321 04:52:07.449743 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:07 crc kubenswrapper[4775]: I0321 04:52:07.468433 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:07 crc kubenswrapper[4775]: I0321 04:52:07.521656 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" podStartSLOduration=13.521634875 podStartE2EDuration="13.521634875s" podCreationTimestamp="2026-03-21 04:51:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:52:07.488276732 +0000 UTC m=+280.464740356" watchObservedRunningTime="2026-03-21 04:52:07.521634875 +0000 UTC m=+280.498098499" Mar 21 04:52:16 crc kubenswrapper[4775]: I0321 04:52:16.271474 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cfvcz" event={"ID":"96069bfd-088a-4053-ab37-76a04683a6a6","Type":"ContainerStarted","Data":"86d840f2a45ca8b3bb3a23817480b80390f3ea92771806965d547cb2a92d185e"} Mar 21 04:52:16 crc kubenswrapper[4775]: I0321 04:52:16.273157 4775 generic.go:334] "Generic (PLEG): container finished" podID="de54a48a-b733-4042-80b6-ecc719712314" containerID="661b6896c2a4aa43857ae9de8bb6472c19fadc3bd2c01a766c40cd5ed12ea550" exitCode=0 Mar 21 04:52:16 crc kubenswrapper[4775]: I0321 04:52:16.273187 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5fgf" event={"ID":"de54a48a-b733-4042-80b6-ecc719712314","Type":"ContainerDied","Data":"661b6896c2a4aa43857ae9de8bb6472c19fadc3bd2c01a766c40cd5ed12ea550"} Mar 21 04:52:17 crc kubenswrapper[4775]: I0321 04:52:17.280972 4775 generic.go:334] "Generic (PLEG): container finished" podID="5bb65dee-cd5f-46b3-9e7d-36e5d182d19e" containerID="8b5a6baf6223e16aada12bd66d033b18e08ee43831676efdf6aa9934129688fc" exitCode=0 Mar 21 04:52:17 crc kubenswrapper[4775]: I0321 04:52:17.281347 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-959kj" event={"ID":"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e","Type":"ContainerDied","Data":"8b5a6baf6223e16aada12bd66d033b18e08ee43831676efdf6aa9934129688fc"} Mar 21 04:52:17 crc kubenswrapper[4775]: I0321 04:52:17.284803 4775 generic.go:334] "Generic (PLEG): container finished" podID="96069bfd-088a-4053-ab37-76a04683a6a6" containerID="86d840f2a45ca8b3bb3a23817480b80390f3ea92771806965d547cb2a92d185e" exitCode=0 Mar 21 04:52:17 crc kubenswrapper[4775]: I0321 04:52:17.284868 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cfvcz" event={"ID":"96069bfd-088a-4053-ab37-76a04683a6a6","Type":"ContainerDied","Data":"86d840f2a45ca8b3bb3a23817480b80390f3ea92771806965d547cb2a92d185e"} Mar 21 04:52:17 crc kubenswrapper[4775]: I0321 04:52:17.288787 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5fgf" event={"ID":"de54a48a-b733-4042-80b6-ecc719712314","Type":"ContainerStarted","Data":"e9e73c129b659ac9d1d7ebb1db8f0b40547803719861caf1d1794c7a799f2a15"} Mar 21 04:52:17 crc kubenswrapper[4775]: I0321 04:52:17.691368 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v5fgf" podStartSLOduration=3.971158257 podStartE2EDuration="1m3.691348032s" podCreationTimestamp="2026-03-21 04:51:14 +0000 UTC" firstStartedPulling="2026-03-21 04:51:16.905813036 +0000 UTC m=+229.882276670" lastFinishedPulling="2026-03-21 04:52:16.626002821 +0000 UTC m=+289.602466445" observedRunningTime="2026-03-21 04:52:17.337404812 +0000 UTC m=+290.313868436" watchObservedRunningTime="2026-03-21 04:52:17.691348032 +0000 UTC m=+290.667811656" Mar 21 04:52:18 crc kubenswrapper[4775]: I0321 04:52:18.303155 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cfvcz" event={"ID":"96069bfd-088a-4053-ab37-76a04683a6a6","Type":"ContainerStarted","Data":"5448b11851da85cc3503470566b1ac0017e53d4c7a81182921503467a33e3be5"} Mar 21 04:52:18 crc kubenswrapper[4775]: I0321 04:52:18.308517 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-959kj" event={"ID":"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e","Type":"ContainerStarted","Data":"aaa815a5f0e220d5cd57bb1c4b415350dcf17e31c1b987c38d4e0714c4280f96"} Mar 21 04:52:18 crc kubenswrapper[4775]: I0321 04:52:18.321654 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cfvcz" podStartSLOduration=8.546578343 podStartE2EDuration="1m1.321636341s" podCreationTimestamp="2026-03-21 04:51:17 +0000 UTC" firstStartedPulling="2026-03-21 04:51:24.993965317 +0000 UTC m=+237.970428941" lastFinishedPulling="2026-03-21 04:52:17.769023315 +0000 UTC m=+290.745486939" observedRunningTime="2026-03-21 04:52:18.321402684 +0000 UTC m=+291.297866318" watchObservedRunningTime="2026-03-21 04:52:18.321636341 +0000 UTC m=+291.298099955" Mar 21 04:52:18 crc kubenswrapper[4775]: I0321 04:52:18.336423 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-959kj" podStartSLOduration=4.503403423 podStartE2EDuration="1m5.336406837s" podCreationTimestamp="2026-03-21 04:51:13 +0000 UTC" firstStartedPulling="2026-03-21 04:51:16.821184955 +0000 UTC m=+229.797648579" lastFinishedPulling="2026-03-21 04:52:17.654188369 +0000 UTC m=+290.630651993" observedRunningTime="2026-03-21 04:52:18.335774399 +0000 UTC m=+291.312238033" watchObservedRunningTime="2026-03-21 04:52:18.336406837 +0000 UTC m=+291.312870461" Mar 21 04:52:19 crc kubenswrapper[4775]: I0321 04:52:19.325500 4775 generic.go:334] "Generic (PLEG): container finished" podID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" containerID="a669beb6a121e654078eda21a56b0d2171a5e04b1968ec341bfe87375574ff9a" exitCode=0 Mar 21 04:52:19 crc kubenswrapper[4775]: I0321 04:52:19.325540 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hldm7" event={"ID":"571e84f2-a2bc-4f09-ac53-d4a4adafa80b","Type":"ContainerDied","Data":"a669beb6a121e654078eda21a56b0d2171a5e04b1968ec341bfe87375574ff9a"} Mar 21 04:52:19 crc kubenswrapper[4775]: I0321 04:52:19.328258 4775 generic.go:334] "Generic (PLEG): container finished" podID="ccb910ab-dcef-4523-81df-c0fb5eb83429" containerID="21eb68b1dc783b9a4153f39746a113d0c901b8fcfacb75d4fa0041622f71eb2d" exitCode=0 Mar 21 04:52:19 crc kubenswrapper[4775]: I0321 04:52:19.328343 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbtgv" event={"ID":"ccb910ab-dcef-4523-81df-c0fb5eb83429","Type":"ContainerDied","Data":"21eb68b1dc783b9a4153f39746a113d0c901b8fcfacb75d4fa0041622f71eb2d"} Mar 21 04:52:24 crc kubenswrapper[4775]: I0321 04:52:24.645155 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-959kj" Mar 21 04:52:24 crc kubenswrapper[4775]: I0321 04:52:24.646261 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-959kj" Mar 21 04:52:24 crc kubenswrapper[4775]: I0321 04:52:24.849079 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v5fgf" Mar 21 04:52:24 crc kubenswrapper[4775]: I0321 04:52:24.849148 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v5fgf" Mar 21 04:52:25 crc kubenswrapper[4775]: I0321 04:52:25.373618 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v5fgf" Mar 21 04:52:25 crc kubenswrapper[4775]: I0321 04:52:25.375772 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-959kj" Mar 21 04:52:25 crc kubenswrapper[4775]: I0321 04:52:25.423254 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-959kj" Mar 21 04:52:25 crc kubenswrapper[4775]: I0321 04:52:25.437751 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v5fgf" Mar 21 04:52:25 crc kubenswrapper[4775]: I0321 04:52:25.890403 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v5fgf"] Mar 21 04:52:27 crc kubenswrapper[4775]: I0321 04:52:27.373298 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v5fgf" podUID="de54a48a-b733-4042-80b6-ecc719712314" containerName="registry-server" containerID="cri-o://e9e73c129b659ac9d1d7ebb1db8f0b40547803719861caf1d1794c7a799f2a15" gracePeriod=2 Mar 21 04:52:27 crc kubenswrapper[4775]: I0321 04:52:27.720949 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cfvcz" Mar 21 04:52:27 crc kubenswrapper[4775]: I0321 04:52:27.721045 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cfvcz" Mar 21 04:52:27 crc kubenswrapper[4775]: I0321 04:52:27.779580 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cfvcz" Mar 21 04:52:28 crc kubenswrapper[4775]: I0321 04:52:28.418789 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cfvcz" Mar 21 04:52:29 crc kubenswrapper[4775]: I0321 04:52:29.385623 4775 generic.go:334] "Generic (PLEG): container finished" podID="de54a48a-b733-4042-80b6-ecc719712314" containerID="e9e73c129b659ac9d1d7ebb1db8f0b40547803719861caf1d1794c7a799f2a15" exitCode=0 Mar 21 04:52:29 crc kubenswrapper[4775]: I0321 04:52:29.385695 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5fgf" event={"ID":"de54a48a-b733-4042-80b6-ecc719712314","Type":"ContainerDied","Data":"e9e73c129b659ac9d1d7ebb1db8f0b40547803719861caf1d1794c7a799f2a15"} Mar 21 04:52:30 crc kubenswrapper[4775]: I0321 04:52:30.095632 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cfvcz"] Mar 21 04:52:30 crc kubenswrapper[4775]: I0321 04:52:30.392141 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cfvcz" podUID="96069bfd-088a-4053-ab37-76a04683a6a6" containerName="registry-server" containerID="cri-o://5448b11851da85cc3503470566b1ac0017e53d4c7a81182921503467a33e3be5" gracePeriod=2 Mar 21 04:52:30 crc kubenswrapper[4775]: I0321 04:52:30.739582 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v5fgf" Mar 21 04:52:30 crc kubenswrapper[4775]: I0321 04:52:30.815560 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de54a48a-b733-4042-80b6-ecc719712314-catalog-content\") pod \"de54a48a-b733-4042-80b6-ecc719712314\" (UID: \"de54a48a-b733-4042-80b6-ecc719712314\") " Mar 21 04:52:30 crc kubenswrapper[4775]: I0321 04:52:30.815639 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98dgw\" (UniqueName: \"kubernetes.io/projected/de54a48a-b733-4042-80b6-ecc719712314-kube-api-access-98dgw\") pod \"de54a48a-b733-4042-80b6-ecc719712314\" (UID: \"de54a48a-b733-4042-80b6-ecc719712314\") " Mar 21 04:52:30 crc kubenswrapper[4775]: I0321 04:52:30.815769 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de54a48a-b733-4042-80b6-ecc719712314-utilities\") pod \"de54a48a-b733-4042-80b6-ecc719712314\" (UID: \"de54a48a-b733-4042-80b6-ecc719712314\") " Mar 21 04:52:30 crc kubenswrapper[4775]: I0321 04:52:30.816726 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de54a48a-b733-4042-80b6-ecc719712314-utilities" (OuterVolumeSpecName: "utilities") pod "de54a48a-b733-4042-80b6-ecc719712314" (UID: "de54a48a-b733-4042-80b6-ecc719712314"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:52:30 crc kubenswrapper[4775]: I0321 04:52:30.823079 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de54a48a-b733-4042-80b6-ecc719712314-kube-api-access-98dgw" (OuterVolumeSpecName: "kube-api-access-98dgw") pod "de54a48a-b733-4042-80b6-ecc719712314" (UID: "de54a48a-b733-4042-80b6-ecc719712314"). InnerVolumeSpecName "kube-api-access-98dgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:52:30 crc kubenswrapper[4775]: I0321 04:52:30.916534 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98dgw\" (UniqueName: \"kubernetes.io/projected/de54a48a-b733-4042-80b6-ecc719712314-kube-api-access-98dgw\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:30 crc kubenswrapper[4775]: I0321 04:52:30.916569 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de54a48a-b733-4042-80b6-ecc719712314-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:31 crc kubenswrapper[4775]: I0321 04:52:31.371271 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de54a48a-b733-4042-80b6-ecc719712314-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de54a48a-b733-4042-80b6-ecc719712314" (UID: "de54a48a-b733-4042-80b6-ecc719712314"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:52:31 crc kubenswrapper[4775]: I0321 04:52:31.401163 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5fgf" event={"ID":"de54a48a-b733-4042-80b6-ecc719712314","Type":"ContainerDied","Data":"ade6fa22510b087e9954c736cf8d4e9716ee7ec2fc194bbd254a73a9267ce630"} Mar 21 04:52:31 crc kubenswrapper[4775]: I0321 04:52:31.401264 4775 scope.go:117] "RemoveContainer" containerID="e9e73c129b659ac9d1d7ebb1db8f0b40547803719861caf1d1794c7a799f2a15" Mar 21 04:52:31 crc kubenswrapper[4775]: I0321 04:52:31.401429 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v5fgf" Mar 21 04:52:31 crc kubenswrapper[4775]: I0321 04:52:31.404009 4775 generic.go:334] "Generic (PLEG): container finished" podID="96069bfd-088a-4053-ab37-76a04683a6a6" containerID="5448b11851da85cc3503470566b1ac0017e53d4c7a81182921503467a33e3be5" exitCode=0 Mar 21 04:52:31 crc kubenswrapper[4775]: I0321 04:52:31.404040 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cfvcz" event={"ID":"96069bfd-088a-4053-ab37-76a04683a6a6","Type":"ContainerDied","Data":"5448b11851da85cc3503470566b1ac0017e53d4c7a81182921503467a33e3be5"} Mar 21 04:52:31 crc kubenswrapper[4775]: I0321 04:52:31.420666 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de54a48a-b733-4042-80b6-ecc719712314-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:31 crc kubenswrapper[4775]: I0321 04:52:31.436639 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v5fgf"] Mar 21 04:52:31 crc kubenswrapper[4775]: I0321 04:52:31.439242 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v5fgf"] Mar 21 04:52:31 crc kubenswrapper[4775]: I0321 04:52:31.626008 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" podUID="fddde3da-8512-4e62-9c38-b59f98e117e0" containerName="oauth-openshift" containerID="cri-o://e7db5d2af2536f63217dcf1801711995965b013eb5a3eabb8f70b3e311b18d76" gracePeriod=15 Mar 21 04:52:31 crc kubenswrapper[4775]: I0321 04:52:31.671014 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de54a48a-b733-4042-80b6-ecc719712314" path="/var/lib/kubelet/pods/de54a48a-b733-4042-80b6-ecc719712314/volumes" Mar 21 04:52:33 crc kubenswrapper[4775]: I0321 04:52:33.211276 4775 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-jd86b container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" start-of-body= Mar 21 04:52:33 crc kubenswrapper[4775]: I0321 04:52:33.211351 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" podUID="fddde3da-8512-4e62-9c38-b59f98e117e0" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" Mar 21 04:52:34 crc kubenswrapper[4775]: I0321 04:52:34.388994 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7bd4667b97-z2cn5"] Mar 21 04:52:34 crc kubenswrapper[4775]: I0321 04:52:34.389531 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" podUID="05791985-4754-4e1e-9400-7204c6a5eab2" containerName="controller-manager" containerID="cri-o://5494d63ffe4e008a81f5adf939d712754549a2efc64457da115e33ba5aea120c" gracePeriod=30 Mar 21 04:52:34 crc kubenswrapper[4775]: I0321 04:52:34.423060 4775 generic.go:334] "Generic (PLEG): container finished" podID="fddde3da-8512-4e62-9c38-b59f98e117e0" containerID="e7db5d2af2536f63217dcf1801711995965b013eb5a3eabb8f70b3e311b18d76" exitCode=0 Mar 21 04:52:34 crc kubenswrapper[4775]: I0321 04:52:34.423107 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" event={"ID":"fddde3da-8512-4e62-9c38-b59f98e117e0","Type":"ContainerDied","Data":"e7db5d2af2536f63217dcf1801711995965b013eb5a3eabb8f70b3e311b18d76"} Mar 21 04:52:34 crc kubenswrapper[4775]: I0321 04:52:34.486925 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8"] Mar 21 04:52:34 crc kubenswrapper[4775]: I0321 04:52:34.487169 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" podUID="ef0e747e-48cd-4a7d-922d-905e15ea750e" containerName="route-controller-manager" containerID="cri-o://2dfd1a63e89a57a5a6969d7c49601a2ff07f69ec16bc071f9034c83d8f03b10b" gracePeriod=30 Mar 21 04:52:35 crc kubenswrapper[4775]: I0321 04:52:35.070928 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cfvcz" Mar 21 04:52:35 crc kubenswrapper[4775]: I0321 04:52:35.077617 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdggw\" (UniqueName: \"kubernetes.io/projected/96069bfd-088a-4053-ab37-76a04683a6a6-kube-api-access-kdggw\") pod \"96069bfd-088a-4053-ab37-76a04683a6a6\" (UID: \"96069bfd-088a-4053-ab37-76a04683a6a6\") " Mar 21 04:52:35 crc kubenswrapper[4775]: I0321 04:52:35.077787 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96069bfd-088a-4053-ab37-76a04683a6a6-utilities\") pod \"96069bfd-088a-4053-ab37-76a04683a6a6\" (UID: \"96069bfd-088a-4053-ab37-76a04683a6a6\") " Mar 21 04:52:35 crc kubenswrapper[4775]: I0321 04:52:35.077902 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96069bfd-088a-4053-ab37-76a04683a6a6-catalog-content\") pod \"96069bfd-088a-4053-ab37-76a04683a6a6\" (UID: \"96069bfd-088a-4053-ab37-76a04683a6a6\") " Mar 21 04:52:35 crc kubenswrapper[4775]: I0321 04:52:35.078767 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96069bfd-088a-4053-ab37-76a04683a6a6-utilities" (OuterVolumeSpecName: "utilities") pod "96069bfd-088a-4053-ab37-76a04683a6a6" (UID: "96069bfd-088a-4053-ab37-76a04683a6a6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:52:35 crc kubenswrapper[4775]: I0321 04:52:35.084220 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96069bfd-088a-4053-ab37-76a04683a6a6-kube-api-access-kdggw" (OuterVolumeSpecName: "kube-api-access-kdggw") pod "96069bfd-088a-4053-ab37-76a04683a6a6" (UID: "96069bfd-088a-4053-ab37-76a04683a6a6"). InnerVolumeSpecName "kube-api-access-kdggw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:52:35 crc kubenswrapper[4775]: I0321 04:52:35.179043 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96069bfd-088a-4053-ab37-76a04683a6a6-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:35 crc kubenswrapper[4775]: I0321 04:52:35.179075 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdggw\" (UniqueName: \"kubernetes.io/projected/96069bfd-088a-4053-ab37-76a04683a6a6-kube-api-access-kdggw\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:35 crc kubenswrapper[4775]: I0321 04:52:35.219948 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96069bfd-088a-4053-ab37-76a04683a6a6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "96069bfd-088a-4053-ab37-76a04683a6a6" (UID: "96069bfd-088a-4053-ab37-76a04683a6a6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:52:35 crc kubenswrapper[4775]: I0321 04:52:35.280553 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96069bfd-088a-4053-ab37-76a04683a6a6-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:35 crc kubenswrapper[4775]: I0321 04:52:35.433104 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cfvcz" event={"ID":"96069bfd-088a-4053-ab37-76a04683a6a6","Type":"ContainerDied","Data":"778fa4db06d864ab1638a1a9d9660eefb0a2a7f44eb1587a31bef24f86e53ae8"} Mar 21 04:52:35 crc kubenswrapper[4775]: I0321 04:52:35.433237 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cfvcz" Mar 21 04:52:35 crc kubenswrapper[4775]: I0321 04:52:35.464912 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cfvcz"] Mar 21 04:52:35 crc kubenswrapper[4775]: I0321 04:52:35.471651 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cfvcz"] Mar 21 04:52:35 crc kubenswrapper[4775]: I0321 04:52:35.667525 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96069bfd-088a-4053-ab37-76a04683a6a6" path="/var/lib/kubelet/pods/96069bfd-088a-4053-ab37-76a04683a6a6/volumes" Mar 21 04:52:36 crc kubenswrapper[4775]: I0321 04:52:36.439926 4775 generic.go:334] "Generic (PLEG): container finished" podID="05791985-4754-4e1e-9400-7204c6a5eab2" containerID="5494d63ffe4e008a81f5adf939d712754549a2efc64457da115e33ba5aea120c" exitCode=0 Mar 21 04:52:36 crc kubenswrapper[4775]: I0321 04:52:36.439982 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" event={"ID":"05791985-4754-4e1e-9400-7204c6a5eab2","Type":"ContainerDied","Data":"5494d63ffe4e008a81f5adf939d712754549a2efc64457da115e33ba5aea120c"} Mar 21 04:52:36 crc kubenswrapper[4775]: I0321 04:52:36.575827 4775 patch_prober.go:28] interesting pod/controller-manager-7bd4667b97-z2cn5 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.62:8443/healthz\": dial tcp 10.217.0.62:8443: connect: connection refused" start-of-body= Mar 21 04:52:36 crc kubenswrapper[4775]: I0321 04:52:36.575887 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" podUID="05791985-4754-4e1e-9400-7204c6a5eab2" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.62:8443/healthz\": dial tcp 10.217.0.62:8443: connect: connection refused" Mar 21 04:52:36 crc kubenswrapper[4775]: I0321 04:52:36.896820 4775 scope.go:117] "RemoveContainer" containerID="661b6896c2a4aa43857ae9de8bb6472c19fadc3bd2c01a766c40cd5ed12ea550" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.215259 4775 scope.go:117] "RemoveContainer" containerID="c42abfff124b04fcc4cb97eb8de832fa3a68e69db6ccb668907d78050544eb6c" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.297509 4775 scope.go:117] "RemoveContainer" containerID="5448b11851da85cc3503470566b1ac0017e53d4c7a81182921503467a33e3be5" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.315787 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.322822 4775 scope.go:117] "RemoveContainer" containerID="86d840f2a45ca8b3bb3a23817480b80390f3ea92771806965d547cb2a92d185e" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.330484 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.376261 4775 scope.go:117] "RemoveContainer" containerID="e2fe2d90884acf37e8aab163e4c967e9d771b2b2c054a404750ee68dc72d022c" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.383879 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.447007 4775 generic.go:334] "Generic (PLEG): container finished" podID="ef0e747e-48cd-4a7d-922d-905e15ea750e" containerID="2dfd1a63e89a57a5a6969d7c49601a2ff07f69ec16bc071f9034c83d8f03b10b" exitCode=0 Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.447210 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" event={"ID":"ef0e747e-48cd-4a7d-922d-905e15ea750e","Type":"ContainerDied","Data":"2dfd1a63e89a57a5a6969d7c49601a2ff07f69ec16bc071f9034c83d8f03b10b"} Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.447999 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" event={"ID":"ef0e747e-48cd-4a7d-922d-905e15ea750e","Type":"ContainerDied","Data":"f1c5e9ed26c65634fec4b7e068160478cde4bffe6efaa83fc8c27834b2f650f2"} Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.447305 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.448030 4775 scope.go:117] "RemoveContainer" containerID="2dfd1a63e89a57a5a6969d7c49601a2ff07f69ec16bc071f9034c83d8f03b10b" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.450861 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.451416 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jd86b" event={"ID":"fddde3da-8512-4e62-9c38-b59f98e117e0","Type":"ContainerDied","Data":"bdc921caf55fdc9483dd16d86e33b3865a46d3e8a029b5b72e32afe0666cef10"} Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.455493 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" event={"ID":"05791985-4754-4e1e-9400-7204c6a5eab2","Type":"ContainerDied","Data":"1b352205937c3361b5c47711682f5ba564f7c8d78431497ed52ef09e4be45784"} Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.455560 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bd4667b97-z2cn5" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.462029 4775 scope.go:117] "RemoveContainer" containerID="2dfd1a63e89a57a5a6969d7c49601a2ff07f69ec16bc071f9034c83d8f03b10b" Mar 21 04:52:37 crc kubenswrapper[4775]: E0321 04:52:37.462411 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dfd1a63e89a57a5a6969d7c49601a2ff07f69ec16bc071f9034c83d8f03b10b\": container with ID starting with 2dfd1a63e89a57a5a6969d7c49601a2ff07f69ec16bc071f9034c83d8f03b10b not found: ID does not exist" containerID="2dfd1a63e89a57a5a6969d7c49601a2ff07f69ec16bc071f9034c83d8f03b10b" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.462447 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dfd1a63e89a57a5a6969d7c49601a2ff07f69ec16bc071f9034c83d8f03b10b"} err="failed to get container status \"2dfd1a63e89a57a5a6969d7c49601a2ff07f69ec16bc071f9034c83d8f03b10b\": rpc error: code = NotFound desc = could not find container \"2dfd1a63e89a57a5a6969d7c49601a2ff07f69ec16bc071f9034c83d8f03b10b\": container with ID starting with 2dfd1a63e89a57a5a6969d7c49601a2ff07f69ec16bc071f9034c83d8f03b10b not found: ID does not exist" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.462471 4775 scope.go:117] "RemoveContainer" containerID="e7db5d2af2536f63217dcf1801711995965b013eb5a3eabb8f70b3e311b18d76" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.495977 4775 scope.go:117] "RemoveContainer" containerID="5494d63ffe4e008a81f5adf939d712754549a2efc64457da115e33ba5aea120c" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.508812 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-template-error\") pod \"fddde3da-8512-4e62-9c38-b59f98e117e0\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.508870 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-router-certs\") pod \"fddde3da-8512-4e62-9c38-b59f98e117e0\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.508899 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-ocp-branding-template\") pod \"fddde3da-8512-4e62-9c38-b59f98e117e0\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.508961 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-template-login\") pod \"fddde3da-8512-4e62-9c38-b59f98e117e0\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.509974 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hc9l\" (UniqueName: \"kubernetes.io/projected/ef0e747e-48cd-4a7d-922d-905e15ea750e-kube-api-access-9hc9l\") pod \"ef0e747e-48cd-4a7d-922d-905e15ea750e\" (UID: \"ef0e747e-48cd-4a7d-922d-905e15ea750e\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.510007 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-cliconfig\") pod \"fddde3da-8512-4e62-9c38-b59f98e117e0\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.510029 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef0e747e-48cd-4a7d-922d-905e15ea750e-config\") pod \"ef0e747e-48cd-4a7d-922d-905e15ea750e\" (UID: \"ef0e747e-48cd-4a7d-922d-905e15ea750e\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.510052 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05791985-4754-4e1e-9400-7204c6a5eab2-client-ca\") pod \"05791985-4754-4e1e-9400-7204c6a5eab2\" (UID: \"05791985-4754-4e1e-9400-7204c6a5eab2\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.510075 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-trusted-ca-bundle\") pod \"fddde3da-8512-4e62-9c38-b59f98e117e0\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.510097 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fw6r\" (UniqueName: \"kubernetes.io/projected/fddde3da-8512-4e62-9c38-b59f98e117e0-kube-api-access-4fw6r\") pod \"fddde3da-8512-4e62-9c38-b59f98e117e0\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.510147 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/05791985-4754-4e1e-9400-7204c6a5eab2-proxy-ca-bundles\") pod \"05791985-4754-4e1e-9400-7204c6a5eab2\" (UID: \"05791985-4754-4e1e-9400-7204c6a5eab2\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.510172 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-idp-0-file-data\") pod \"fddde3da-8512-4e62-9c38-b59f98e117e0\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.510193 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ef0e747e-48cd-4a7d-922d-905e15ea750e-client-ca\") pod \"ef0e747e-48cd-4a7d-922d-905e15ea750e\" (UID: \"ef0e747e-48cd-4a7d-922d-905e15ea750e\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.510221 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-template-provider-selection\") pod \"fddde3da-8512-4e62-9c38-b59f98e117e0\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.510247 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef0e747e-48cd-4a7d-922d-905e15ea750e-serving-cert\") pod \"ef0e747e-48cd-4a7d-922d-905e15ea750e\" (UID: \"ef0e747e-48cd-4a7d-922d-905e15ea750e\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.510277 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05791985-4754-4e1e-9400-7204c6a5eab2-config\") pod \"05791985-4754-4e1e-9400-7204c6a5eab2\" (UID: \"05791985-4754-4e1e-9400-7204c6a5eab2\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.510300 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-serving-cert\") pod \"fddde3da-8512-4e62-9c38-b59f98e117e0\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.510320 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05791985-4754-4e1e-9400-7204c6a5eab2-serving-cert\") pod \"05791985-4754-4e1e-9400-7204c6a5eab2\" (UID: \"05791985-4754-4e1e-9400-7204c6a5eab2\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.510420 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-session\") pod \"fddde3da-8512-4e62-9c38-b59f98e117e0\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.510450 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-audit-policies\") pod \"fddde3da-8512-4e62-9c38-b59f98e117e0\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.510473 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6s67\" (UniqueName: \"kubernetes.io/projected/05791985-4754-4e1e-9400-7204c6a5eab2-kube-api-access-l6s67\") pod \"05791985-4754-4e1e-9400-7204c6a5eab2\" (UID: \"05791985-4754-4e1e-9400-7204c6a5eab2\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.510506 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-service-ca\") pod \"fddde3da-8512-4e62-9c38-b59f98e117e0\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.510527 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fddde3da-8512-4e62-9c38-b59f98e117e0-audit-dir\") pod \"fddde3da-8512-4e62-9c38-b59f98e117e0\" (UID: \"fddde3da-8512-4e62-9c38-b59f98e117e0\") " Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.510765 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fddde3da-8512-4e62-9c38-b59f98e117e0-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "fddde3da-8512-4e62-9c38-b59f98e117e0" (UID: "fddde3da-8512-4e62-9c38-b59f98e117e0"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.511870 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "fddde3da-8512-4e62-9c38-b59f98e117e0" (UID: "fddde3da-8512-4e62-9c38-b59f98e117e0"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.512335 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05791985-4754-4e1e-9400-7204c6a5eab2-client-ca" (OuterVolumeSpecName: "client-ca") pod "05791985-4754-4e1e-9400-7204c6a5eab2" (UID: "05791985-4754-4e1e-9400-7204c6a5eab2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.512500 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef0e747e-48cd-4a7d-922d-905e15ea750e-config" (OuterVolumeSpecName: "config") pod "ef0e747e-48cd-4a7d-922d-905e15ea750e" (UID: "ef0e747e-48cd-4a7d-922d-905e15ea750e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.512522 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05791985-4754-4e1e-9400-7204c6a5eab2-config" (OuterVolumeSpecName: "config") pod "05791985-4754-4e1e-9400-7204c6a5eab2" (UID: "05791985-4754-4e1e-9400-7204c6a5eab2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.513207 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05791985-4754-4e1e-9400-7204c6a5eab2-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "05791985-4754-4e1e-9400-7204c6a5eab2" (UID: "05791985-4754-4e1e-9400-7204c6a5eab2"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.513624 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05791985-4754-4e1e-9400-7204c6a5eab2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "05791985-4754-4e1e-9400-7204c6a5eab2" (UID: "05791985-4754-4e1e-9400-7204c6a5eab2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.514040 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "fddde3da-8512-4e62-9c38-b59f98e117e0" (UID: "fddde3da-8512-4e62-9c38-b59f98e117e0"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.515359 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef0e747e-48cd-4a7d-922d-905e15ea750e-kube-api-access-9hc9l" (OuterVolumeSpecName: "kube-api-access-9hc9l") pod "ef0e747e-48cd-4a7d-922d-905e15ea750e" (UID: "ef0e747e-48cd-4a7d-922d-905e15ea750e"). InnerVolumeSpecName "kube-api-access-9hc9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.515704 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "fddde3da-8512-4e62-9c38-b59f98e117e0" (UID: "fddde3da-8512-4e62-9c38-b59f98e117e0"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.515927 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "fddde3da-8512-4e62-9c38-b59f98e117e0" (UID: "fddde3da-8512-4e62-9c38-b59f98e117e0"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.516085 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef0e747e-48cd-4a7d-922d-905e15ea750e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ef0e747e-48cd-4a7d-922d-905e15ea750e" (UID: "ef0e747e-48cd-4a7d-922d-905e15ea750e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.516328 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05791985-4754-4e1e-9400-7204c6a5eab2-kube-api-access-l6s67" (OuterVolumeSpecName: "kube-api-access-l6s67") pod "05791985-4754-4e1e-9400-7204c6a5eab2" (UID: "05791985-4754-4e1e-9400-7204c6a5eab2"). InnerVolumeSpecName "kube-api-access-l6s67". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.516483 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "fddde3da-8512-4e62-9c38-b59f98e117e0" (UID: "fddde3da-8512-4e62-9c38-b59f98e117e0"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.516670 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "fddde3da-8512-4e62-9c38-b59f98e117e0" (UID: "fddde3da-8512-4e62-9c38-b59f98e117e0"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.517019 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "fddde3da-8512-4e62-9c38-b59f98e117e0" (UID: "fddde3da-8512-4e62-9c38-b59f98e117e0"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.517277 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef0e747e-48cd-4a7d-922d-905e15ea750e-client-ca" (OuterVolumeSpecName: "client-ca") pod "ef0e747e-48cd-4a7d-922d-905e15ea750e" (UID: "ef0e747e-48cd-4a7d-922d-905e15ea750e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.518105 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "fddde3da-8512-4e62-9c38-b59f98e117e0" (UID: "fddde3da-8512-4e62-9c38-b59f98e117e0"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.519825 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "fddde3da-8512-4e62-9c38-b59f98e117e0" (UID: "fddde3da-8512-4e62-9c38-b59f98e117e0"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.520505 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fddde3da-8512-4e62-9c38-b59f98e117e0-kube-api-access-4fw6r" (OuterVolumeSpecName: "kube-api-access-4fw6r") pod "fddde3da-8512-4e62-9c38-b59f98e117e0" (UID: "fddde3da-8512-4e62-9c38-b59f98e117e0"). InnerVolumeSpecName "kube-api-access-4fw6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.520651 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "fddde3da-8512-4e62-9c38-b59f98e117e0" (UID: "fddde3da-8512-4e62-9c38-b59f98e117e0"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.520822 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "fddde3da-8512-4e62-9c38-b59f98e117e0" (UID: "fddde3da-8512-4e62-9c38-b59f98e117e0"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.523621 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "fddde3da-8512-4e62-9c38-b59f98e117e0" (UID: "fddde3da-8512-4e62-9c38-b59f98e117e0"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611299 4775 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611336 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6s67\" (UniqueName: \"kubernetes.io/projected/05791985-4754-4e1e-9400-7204c6a5eab2-kube-api-access-l6s67\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611352 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611363 4775 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fddde3da-8512-4e62-9c38-b59f98e117e0-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611374 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611385 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611396 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611408 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611419 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hc9l\" (UniqueName: \"kubernetes.io/projected/ef0e747e-48cd-4a7d-922d-905e15ea750e-kube-api-access-9hc9l\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611430 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611440 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef0e747e-48cd-4a7d-922d-905e15ea750e-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611452 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05791985-4754-4e1e-9400-7204c6a5eab2-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611462 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611473 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fw6r\" (UniqueName: \"kubernetes.io/projected/fddde3da-8512-4e62-9c38-b59f98e117e0-kube-api-access-4fw6r\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611484 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/05791985-4754-4e1e-9400-7204c6a5eab2-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611494 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611504 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ef0e747e-48cd-4a7d-922d-905e15ea750e-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611517 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611527 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef0e747e-48cd-4a7d-922d-905e15ea750e-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611537 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05791985-4754-4e1e-9400-7204c6a5eab2-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611546 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611561 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05791985-4754-4e1e-9400-7204c6a5eab2-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.611573 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fddde3da-8512-4e62-9c38-b59f98e117e0-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.800673 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8"] Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.823437 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fbcc6866-hpvz8"] Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.829876 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jd86b"] Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.839655 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jd86b"] Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.844109 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7bd4667b97-z2cn5"] Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.847753 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7bd4667b97-z2cn5"] Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.977444 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6f88f547f-kjxrc"] Mar 21 04:52:37 crc kubenswrapper[4775]: E0321 04:52:37.977883 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de54a48a-b733-4042-80b6-ecc719712314" containerName="extract-utilities" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.977903 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="de54a48a-b733-4042-80b6-ecc719712314" containerName="extract-utilities" Mar 21 04:52:37 crc kubenswrapper[4775]: E0321 04:52:37.977916 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96069bfd-088a-4053-ab37-76a04683a6a6" containerName="extract-utilities" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.977923 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="96069bfd-088a-4053-ab37-76a04683a6a6" containerName="extract-utilities" Mar 21 04:52:37 crc kubenswrapper[4775]: E0321 04:52:37.977934 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fddde3da-8512-4e62-9c38-b59f98e117e0" containerName="oauth-openshift" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.977943 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fddde3da-8512-4e62-9c38-b59f98e117e0" containerName="oauth-openshift" Mar 21 04:52:37 crc kubenswrapper[4775]: E0321 04:52:37.977954 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de54a48a-b733-4042-80b6-ecc719712314" containerName="extract-content" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.977962 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="de54a48a-b733-4042-80b6-ecc719712314" containerName="extract-content" Mar 21 04:52:37 crc kubenswrapper[4775]: E0321 04:52:37.977979 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6977ba5-d878-4457-a6d4-8acd42ebb089" containerName="pruner" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.977987 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6977ba5-d878-4457-a6d4-8acd42ebb089" containerName="pruner" Mar 21 04:52:37 crc kubenswrapper[4775]: E0321 04:52:37.978005 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8101654-10fd-404b-ae0f-a098719418f4" containerName="oc" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.978012 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8101654-10fd-404b-ae0f-a098719418f4" containerName="oc" Mar 21 04:52:37 crc kubenswrapper[4775]: E0321 04:52:37.978026 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96069bfd-088a-4053-ab37-76a04683a6a6" containerName="registry-server" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.978035 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="96069bfd-088a-4053-ab37-76a04683a6a6" containerName="registry-server" Mar 21 04:52:37 crc kubenswrapper[4775]: E0321 04:52:37.978049 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96069bfd-088a-4053-ab37-76a04683a6a6" containerName="extract-content" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.978056 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="96069bfd-088a-4053-ab37-76a04683a6a6" containerName="extract-content" Mar 21 04:52:37 crc kubenswrapper[4775]: E0321 04:52:37.978065 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de54a48a-b733-4042-80b6-ecc719712314" containerName="registry-server" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.978072 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="de54a48a-b733-4042-80b6-ecc719712314" containerName="registry-server" Mar 21 04:52:37 crc kubenswrapper[4775]: E0321 04:52:37.978084 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05791985-4754-4e1e-9400-7204c6a5eab2" containerName="controller-manager" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.978091 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="05791985-4754-4e1e-9400-7204c6a5eab2" containerName="controller-manager" Mar 21 04:52:37 crc kubenswrapper[4775]: E0321 04:52:37.978099 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef0e747e-48cd-4a7d-922d-905e15ea750e" containerName="route-controller-manager" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.978108 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef0e747e-48cd-4a7d-922d-905e15ea750e" containerName="route-controller-manager" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.978239 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="05791985-4754-4e1e-9400-7204c6a5eab2" containerName="controller-manager" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.978252 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6977ba5-d878-4457-a6d4-8acd42ebb089" containerName="pruner" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.978263 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="96069bfd-088a-4053-ab37-76a04683a6a6" containerName="registry-server" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.978275 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef0e747e-48cd-4a7d-922d-905e15ea750e" containerName="route-controller-manager" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.978287 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fddde3da-8512-4e62-9c38-b59f98e117e0" containerName="oauth-openshift" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.978309 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8101654-10fd-404b-ae0f-a098719418f4" containerName="oc" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.978321 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="de54a48a-b733-4042-80b6-ecc719712314" containerName="registry-server" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.978742 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.981039 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.981390 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.981581 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.982070 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.982279 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.983034 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.994649 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 21 04:52:37 crc kubenswrapper[4775]: I0321 04:52:37.996673 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6f88f547f-kjxrc"] Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.022867 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7da7ce8b-402a-40f7-a9ce-340524ad8573-proxy-ca-bundles\") pod \"controller-manager-6f88f547f-kjxrc\" (UID: \"7da7ce8b-402a-40f7-a9ce-340524ad8573\") " pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.022914 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7da7ce8b-402a-40f7-a9ce-340524ad8573-client-ca\") pod \"controller-manager-6f88f547f-kjxrc\" (UID: \"7da7ce8b-402a-40f7-a9ce-340524ad8573\") " pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.022942 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7da7ce8b-402a-40f7-a9ce-340524ad8573-serving-cert\") pod \"controller-manager-6f88f547f-kjxrc\" (UID: \"7da7ce8b-402a-40f7-a9ce-340524ad8573\") " pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.022984 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flbhf\" (UniqueName: \"kubernetes.io/projected/7da7ce8b-402a-40f7-a9ce-340524ad8573-kube-api-access-flbhf\") pod \"controller-manager-6f88f547f-kjxrc\" (UID: \"7da7ce8b-402a-40f7-a9ce-340524ad8573\") " pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.023017 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da7ce8b-402a-40f7-a9ce-340524ad8573-config\") pod \"controller-manager-6f88f547f-kjxrc\" (UID: \"7da7ce8b-402a-40f7-a9ce-340524ad8573\") " pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.124107 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da7ce8b-402a-40f7-a9ce-340524ad8573-config\") pod \"controller-manager-6f88f547f-kjxrc\" (UID: \"7da7ce8b-402a-40f7-a9ce-340524ad8573\") " pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.124185 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7da7ce8b-402a-40f7-a9ce-340524ad8573-proxy-ca-bundles\") pod \"controller-manager-6f88f547f-kjxrc\" (UID: \"7da7ce8b-402a-40f7-a9ce-340524ad8573\") " pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.124209 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7da7ce8b-402a-40f7-a9ce-340524ad8573-client-ca\") pod \"controller-manager-6f88f547f-kjxrc\" (UID: \"7da7ce8b-402a-40f7-a9ce-340524ad8573\") " pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.124243 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7da7ce8b-402a-40f7-a9ce-340524ad8573-serving-cert\") pod \"controller-manager-6f88f547f-kjxrc\" (UID: \"7da7ce8b-402a-40f7-a9ce-340524ad8573\") " pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.124284 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flbhf\" (UniqueName: \"kubernetes.io/projected/7da7ce8b-402a-40f7-a9ce-340524ad8573-kube-api-access-flbhf\") pod \"controller-manager-6f88f547f-kjxrc\" (UID: \"7da7ce8b-402a-40f7-a9ce-340524ad8573\") " pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.125529 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7da7ce8b-402a-40f7-a9ce-340524ad8573-proxy-ca-bundles\") pod \"controller-manager-6f88f547f-kjxrc\" (UID: \"7da7ce8b-402a-40f7-a9ce-340524ad8573\") " pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.125667 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7da7ce8b-402a-40f7-a9ce-340524ad8573-client-ca\") pod \"controller-manager-6f88f547f-kjxrc\" (UID: \"7da7ce8b-402a-40f7-a9ce-340524ad8573\") " pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.126229 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da7ce8b-402a-40f7-a9ce-340524ad8573-config\") pod \"controller-manager-6f88f547f-kjxrc\" (UID: \"7da7ce8b-402a-40f7-a9ce-340524ad8573\") " pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.128284 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7da7ce8b-402a-40f7-a9ce-340524ad8573-serving-cert\") pod \"controller-manager-6f88f547f-kjxrc\" (UID: \"7da7ce8b-402a-40f7-a9ce-340524ad8573\") " pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.140818 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flbhf\" (UniqueName: \"kubernetes.io/projected/7da7ce8b-402a-40f7-a9ce-340524ad8573-kube-api-access-flbhf\") pod \"controller-manager-6f88f547f-kjxrc\" (UID: \"7da7ce8b-402a-40f7-a9ce-340524ad8573\") " pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.293345 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.476241 4775 generic.go:334] "Generic (PLEG): container finished" podID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" containerID="a41e12fdcec460e076316274065e17d2ca1fc9dae7706fccd22acc387dfd57b1" exitCode=0 Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.476627 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svf9w" event={"ID":"1dcbca72-150a-47c6-ac3c-f701ae82e05b","Type":"ContainerDied","Data":"a41e12fdcec460e076316274065e17d2ca1fc9dae7706fccd22acc387dfd57b1"} Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.483583 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hldm7" event={"ID":"571e84f2-a2bc-4f09-ac53-d4a4adafa80b","Type":"ContainerStarted","Data":"bce5cf8972de39d9de712cead6829664ca56252c15f4837eb08ebb6d9c3a039b"} Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.488057 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbtgv" event={"ID":"ccb910ab-dcef-4523-81df-c0fb5eb83429","Type":"ContainerStarted","Data":"02ab090064cf43714a361f98c8401a08ff5684170017c160d3affd080275fc5a"} Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.493905 4775 generic.go:334] "Generic (PLEG): container finished" podID="70ad413a-5f81-4094-b2d8-9b89698c6e32" containerID="1098f33834501f7272dae8532213ab991de93513a5cd5e8028551095da4598b0" exitCode=0 Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.493998 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjfwd" event={"ID":"70ad413a-5f81-4094-b2d8-9b89698c6e32","Type":"ContainerDied","Data":"1098f33834501f7272dae8532213ab991de93513a5cd5e8028551095da4598b0"} Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.497897 4775 generic.go:334] "Generic (PLEG): container finished" podID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" containerID="93e802bba3a0a60860befd9b343fb138a5913ba96b3cd2219f5f37eec14db557" exitCode=0 Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.497943 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7k9df" event={"ID":"55e2733e-3620-4ecb-a51e-33b1fd3dce9d","Type":"ContainerDied","Data":"93e802bba3a0a60860befd9b343fb138a5913ba96b3cd2219f5f37eec14db557"} Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.554980 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wbtgv" podStartSLOduration=6.418065519 podStartE2EDuration="1m24.55496296s" podCreationTimestamp="2026-03-21 04:51:14 +0000 UTC" firstStartedPulling="2026-03-21 04:51:16.892010759 +0000 UTC m=+229.868474373" lastFinishedPulling="2026-03-21 04:52:35.02890819 +0000 UTC m=+308.005371814" observedRunningTime="2026-03-21 04:52:38.554581279 +0000 UTC m=+311.531044913" watchObservedRunningTime="2026-03-21 04:52:38.55496296 +0000 UTC m=+311.531426584" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.574915 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hldm7" podStartSLOduration=6.632767548 podStartE2EDuration="1m22.574894485s" podCreationTimestamp="2026-03-21 04:51:16 +0000 UTC" firstStartedPulling="2026-03-21 04:51:20.874312246 +0000 UTC m=+233.850775870" lastFinishedPulling="2026-03-21 04:52:36.816439173 +0000 UTC m=+309.792902807" observedRunningTime="2026-03-21 04:52:38.572748593 +0000 UTC m=+311.549212227" watchObservedRunningTime="2026-03-21 04:52:38.574894485 +0000 UTC m=+311.551358109" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.696885 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6f88f547f-kjxrc"] Mar 21 04:52:38 crc kubenswrapper[4775]: W0321 04:52:38.701482 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7da7ce8b_402a_40f7_a9ce_340524ad8573.slice/crio-2e2beab81a4014c294e6052e86efc7d9c966045a25b61fae9abde1855c0e91eb WatchSource:0}: Error finding container 2e2beab81a4014c294e6052e86efc7d9c966045a25b61fae9abde1855c0e91eb: Status 404 returned error can't find the container with id 2e2beab81a4014c294e6052e86efc7d9c966045a25b61fae9abde1855c0e91eb Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.975751 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z"] Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.976891 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.980182 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.980283 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.980720 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.980935 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.981187 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.981244 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 21 04:52:38 crc kubenswrapper[4775]: I0321 04:52:38.992241 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z"] Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.137724 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1882b1c5-2465-40df-9306-e5792e0d9f2f-config\") pod \"route-controller-manager-6d6c87d469-rgj5z\" (UID: \"1882b1c5-2465-40df-9306-e5792e0d9f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.137771 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1882b1c5-2465-40df-9306-e5792e0d9f2f-client-ca\") pod \"route-controller-manager-6d6c87d469-rgj5z\" (UID: \"1882b1c5-2465-40df-9306-e5792e0d9f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.137877 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65v5z\" (UniqueName: \"kubernetes.io/projected/1882b1c5-2465-40df-9306-e5792e0d9f2f-kube-api-access-65v5z\") pod \"route-controller-manager-6d6c87d469-rgj5z\" (UID: \"1882b1c5-2465-40df-9306-e5792e0d9f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.138224 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1882b1c5-2465-40df-9306-e5792e0d9f2f-serving-cert\") pod \"route-controller-manager-6d6c87d469-rgj5z\" (UID: \"1882b1c5-2465-40df-9306-e5792e0d9f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.239628 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1882b1c5-2465-40df-9306-e5792e0d9f2f-serving-cert\") pod \"route-controller-manager-6d6c87d469-rgj5z\" (UID: \"1882b1c5-2465-40df-9306-e5792e0d9f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.240191 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1882b1c5-2465-40df-9306-e5792e0d9f2f-config\") pod \"route-controller-manager-6d6c87d469-rgj5z\" (UID: \"1882b1c5-2465-40df-9306-e5792e0d9f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.240226 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1882b1c5-2465-40df-9306-e5792e0d9f2f-client-ca\") pod \"route-controller-manager-6d6c87d469-rgj5z\" (UID: \"1882b1c5-2465-40df-9306-e5792e0d9f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.240378 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65v5z\" (UniqueName: \"kubernetes.io/projected/1882b1c5-2465-40df-9306-e5792e0d9f2f-kube-api-access-65v5z\") pod \"route-controller-manager-6d6c87d469-rgj5z\" (UID: \"1882b1c5-2465-40df-9306-e5792e0d9f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.241249 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1882b1c5-2465-40df-9306-e5792e0d9f2f-client-ca\") pod \"route-controller-manager-6d6c87d469-rgj5z\" (UID: \"1882b1c5-2465-40df-9306-e5792e0d9f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.241616 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1882b1c5-2465-40df-9306-e5792e0d9f2f-config\") pod \"route-controller-manager-6d6c87d469-rgj5z\" (UID: \"1882b1c5-2465-40df-9306-e5792e0d9f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.245634 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1882b1c5-2465-40df-9306-e5792e0d9f2f-serving-cert\") pod \"route-controller-manager-6d6c87d469-rgj5z\" (UID: \"1882b1c5-2465-40df-9306-e5792e0d9f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.261182 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65v5z\" (UniqueName: \"kubernetes.io/projected/1882b1c5-2465-40df-9306-e5792e0d9f2f-kube-api-access-65v5z\") pod \"route-controller-manager-6d6c87d469-rgj5z\" (UID: \"1882b1c5-2465-40df-9306-e5792e0d9f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.325136 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.515249 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svf9w" event={"ID":"1dcbca72-150a-47c6-ac3c-f701ae82e05b","Type":"ContainerStarted","Data":"924b8554190993f0392bd8c20440e982df121b27ea4559dd16429c3faaecbb7a"} Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.517892 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" event={"ID":"7da7ce8b-402a-40f7-a9ce-340524ad8573","Type":"ContainerStarted","Data":"b57b29e11c242340693c81affc2d61431477f8ba75c091165a111504297edd4f"} Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.517926 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" event={"ID":"7da7ce8b-402a-40f7-a9ce-340524ad8573","Type":"ContainerStarted","Data":"2e2beab81a4014c294e6052e86efc7d9c966045a25b61fae9abde1855c0e91eb"} Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.518144 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.521772 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjfwd" event={"ID":"70ad413a-5f81-4094-b2d8-9b89698c6e32","Type":"ContainerStarted","Data":"dd9cc974af699ec57297810c7016806ecad201037353fbcf0dc303241f78fc0b"} Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.525977 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.532554 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7k9df" event={"ID":"55e2733e-3620-4ecb-a51e-33b1fd3dce9d","Type":"ContainerStarted","Data":"e5e47a7de8c354ec132dd0116fce768c5a764f807e5342721397e5aed3144402"} Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.535696 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-svf9w" podStartSLOduration=3.283391472 podStartE2EDuration="1m25.535684827s" podCreationTimestamp="2026-03-21 04:51:14 +0000 UTC" firstStartedPulling="2026-03-21 04:51:16.879404137 +0000 UTC m=+229.855867761" lastFinishedPulling="2026-03-21 04:52:39.131697492 +0000 UTC m=+312.108161116" observedRunningTime="2026-03-21 04:52:39.533670869 +0000 UTC m=+312.510134493" watchObservedRunningTime="2026-03-21 04:52:39.535684827 +0000 UTC m=+312.512148451" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.572009 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" podStartSLOduration=5.571987155 podStartE2EDuration="5.571987155s" podCreationTimestamp="2026-03-21 04:52:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:52:39.55726094 +0000 UTC m=+312.533724564" watchObservedRunningTime="2026-03-21 04:52:39.571987155 +0000 UTC m=+312.548450779" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.606877 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7k9df" podStartSLOduration=9.692104696 podStartE2EDuration="1m23.606859242s" podCreationTimestamp="2026-03-21 04:51:16 +0000 UTC" firstStartedPulling="2026-03-21 04:51:24.991229569 +0000 UTC m=+237.967693193" lastFinishedPulling="2026-03-21 04:52:38.905984115 +0000 UTC m=+311.882447739" observedRunningTime="2026-03-21 04:52:39.605166223 +0000 UTC m=+312.581629847" watchObservedRunningTime="2026-03-21 04:52:39.606859242 +0000 UTC m=+312.583322866" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.607772 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wjfwd" podStartSLOduration=8.712020172 podStartE2EDuration="1m22.607763008s" podCreationTimestamp="2026-03-21 04:51:17 +0000 UTC" firstStartedPulling="2026-03-21 04:51:25.042959754 +0000 UTC m=+238.019423378" lastFinishedPulling="2026-03-21 04:52:38.93870259 +0000 UTC m=+311.915166214" observedRunningTime="2026-03-21 04:52:39.579462211 +0000 UTC m=+312.555925835" watchObservedRunningTime="2026-03-21 04:52:39.607763008 +0000 UTC m=+312.584226632" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.670761 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05791985-4754-4e1e-9400-7204c6a5eab2" path="/var/lib/kubelet/pods/05791985-4754-4e1e-9400-7204c6a5eab2/volumes" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.671609 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef0e747e-48cd-4a7d-922d-905e15ea750e" path="/var/lib/kubelet/pods/ef0e747e-48cd-4a7d-922d-905e15ea750e/volumes" Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.672240 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fddde3da-8512-4e62-9c38-b59f98e117e0" path="/var/lib/kubelet/pods/fddde3da-8512-4e62-9c38-b59f98e117e0/volumes" Mar 21 04:52:39 crc kubenswrapper[4775]: W0321 04:52:39.736665 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1882b1c5_2465_40df_9306_e5792e0d9f2f.slice/crio-1a6793f0695cc3d26a11dee78d3b1e2f638d464d822526dcc3f7aa8f363599f1 WatchSource:0}: Error finding container 1a6793f0695cc3d26a11dee78d3b1e2f638d464d822526dcc3f7aa8f363599f1: Status 404 returned error can't find the container with id 1a6793f0695cc3d26a11dee78d3b1e2f638d464d822526dcc3f7aa8f363599f1 Mar 21 04:52:39 crc kubenswrapper[4775]: I0321 04:52:39.737323 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z"] Mar 21 04:52:40 crc kubenswrapper[4775]: I0321 04:52:40.550336 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" event={"ID":"1882b1c5-2465-40df-9306-e5792e0d9f2f","Type":"ContainerStarted","Data":"ae81d0a5c5add9cc57df7dd09c08b1e07e8506e8be71c91d0656558f8529bd3b"} Mar 21 04:52:40 crc kubenswrapper[4775]: I0321 04:52:40.550394 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" event={"ID":"1882b1c5-2465-40df-9306-e5792e0d9f2f","Type":"ContainerStarted","Data":"1a6793f0695cc3d26a11dee78d3b1e2f638d464d822526dcc3f7aa8f363599f1"} Mar 21 04:52:40 crc kubenswrapper[4775]: I0321 04:52:40.550887 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" Mar 21 04:52:40 crc kubenswrapper[4775]: I0321 04:52:40.555527 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" Mar 21 04:52:40 crc kubenswrapper[4775]: I0321 04:52:40.568816 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" podStartSLOduration=6.568799767 podStartE2EDuration="6.568799767s" podCreationTimestamp="2026-03-21 04:52:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:52:40.565519792 +0000 UTC m=+313.541983416" watchObservedRunningTime="2026-03-21 04:52:40.568799767 +0000 UTC m=+313.545263391" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.901471 4775 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.902212 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.902474 4775 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.903020 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70" gracePeriod=15 Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.903063 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401" gracePeriod=15 Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.903100 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc" gracePeriod=15 Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.903100 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb" gracePeriod=15 Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.903252 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948" gracePeriod=15 Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.904578 4775 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 21 04:52:41 crc kubenswrapper[4775]: E0321 04:52:41.904715 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.904731 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 04:52:41 crc kubenswrapper[4775]: E0321 04:52:41.904738 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.904744 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 21 04:52:41 crc kubenswrapper[4775]: E0321 04:52:41.904750 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.904756 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 04:52:41 crc kubenswrapper[4775]: E0321 04:52:41.904766 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.904772 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 21 04:52:41 crc kubenswrapper[4775]: E0321 04:52:41.904782 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.904788 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 04:52:41 crc kubenswrapper[4775]: E0321 04:52:41.904795 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.904800 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 21 04:52:41 crc kubenswrapper[4775]: E0321 04:52:41.904808 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.904814 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 21 04:52:41 crc kubenswrapper[4775]: E0321 04:52:41.904821 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.904827 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.904938 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.904947 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.904954 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.904963 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.904989 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.905000 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.905009 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 21 04:52:41 crc kubenswrapper[4775]: E0321 04:52:41.906110 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.906143 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.906259 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.906268 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 04:52:41 crc kubenswrapper[4775]: E0321 04:52:41.906362 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.906370 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.960655 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.981640 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.981726 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.981753 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.981774 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.981792 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.981807 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.981836 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:52:41 crc kubenswrapper[4775]: I0321 04:52:41.981851 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.085673 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.085746 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.085766 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.085778 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.085819 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.085825 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.085839 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.085859 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.085912 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.085922 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.086017 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.086026 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.086061 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.086077 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.085916 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.086164 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.258924 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 04:52:42 crc kubenswrapper[4775]: W0321 04:52:42.285713 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-4e3fc3de5b7765ff18c8282fe983ce2daa4b942f0497b52cf4a498cd4d684581 WatchSource:0}: Error finding container 4e3fc3de5b7765ff18c8282fe983ce2daa4b942f0497b52cf4a498cd4d684581: Status 404 returned error can't find the container with id 4e3fc3de5b7765ff18c8282fe983ce2daa4b942f0497b52cf4a498cd4d684581 Mar 21 04:52:42 crc kubenswrapper[4775]: E0321 04:52:42.289301 4775 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.200:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189ec223cf4ca9d3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:52:42.288105939 +0000 UTC m=+315.264569563,LastTimestamp:2026-03-21 04:52:42.288105939 +0000 UTC m=+315.264569563,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.568038 4775 generic.go:334] "Generic (PLEG): container finished" podID="bc0e122c-1775-4bf1-9025-6288c383b3f2" containerID="de0bb723a8e5a2f1ecac5f9d70f09000ecc4cbf0b2ce82838fe0f1932e47f097" exitCode=0 Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.568157 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bc0e122c-1775-4bf1-9025-6288c383b3f2","Type":"ContainerDied","Data":"de0bb723a8e5a2f1ecac5f9d70f09000ecc4cbf0b2ce82838fe0f1932e47f097"} Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.572349 4775 status_manager.go:851] "Failed to get status for pod" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.572677 4775 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.572929 4775 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.577214 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.578662 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.579255 4775 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401" exitCode=0 Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.579277 4775 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948" exitCode=0 Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.579286 4775 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb" exitCode=0 Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.579294 4775 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc" exitCode=2 Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.579370 4775 scope.go:117] "RemoveContainer" containerID="85e412eedce18f7222c49be795b325ffdb05306df25f8ff195a08162142324b4" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.582299 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"4e3fc3de5b7765ff18c8282fe983ce2daa4b942f0497b52cf4a498cd4d684581"} Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.583191 4775 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.583630 4775 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.584009 4775 status_manager.go:851] "Failed to get status for pod" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:42 crc kubenswrapper[4775]: E0321 04:52:42.736927 4775 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:42 crc kubenswrapper[4775]: E0321 04:52:42.737392 4775 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:42 crc kubenswrapper[4775]: E0321 04:52:42.737817 4775 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:42 crc kubenswrapper[4775]: E0321 04:52:42.738111 4775 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:42 crc kubenswrapper[4775]: E0321 04:52:42.738433 4775 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:42 crc kubenswrapper[4775]: I0321 04:52:42.738480 4775 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 21 04:52:42 crc kubenswrapper[4775]: E0321 04:52:42.738780 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="200ms" Mar 21 04:52:42 crc kubenswrapper[4775]: E0321 04:52:42.939894 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="400ms" Mar 21 04:52:43 crc kubenswrapper[4775]: I0321 04:52:43.133995 4775 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Mar 21 04:52:43 crc kubenswrapper[4775]: I0321 04:52:43.134091 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Mar 21 04:52:43 crc kubenswrapper[4775]: E0321 04:52:43.227181 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:52:43Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:52:43Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:52:43Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:52:43Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4869c69128f74d9c3b178ea6c8c8d38df169b6bce05eb821a65f0aaf514c563a\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:f3c2ad90e251062165f8d6623ca4994c0b3e28324e4b5b17fd588b162ec97766\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1746912226},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:04dd6a6688563ae1dbeb924643a49d81150c900e2f7e81aaab8a3586e6f53044\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:bc9a28a1c3a5f16a1873e4919f236a9824c62c6394093adb4850b30daf95761c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1252939492},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:898c67bf7fc973e99114f3148976a6c21ae0dbe413051415588fa9b995f5b331\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:a641939d2096609a4cf6eec872a1476b7c671bfd81cffc2edeb6e9f13c9deeba\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1231028434},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0634296304362d86f6b8d0de6c3e1b327670f86a3e0a664a7fd2fa3984fbbb18\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:1a955538926dc95f0f30489b33552cf1f18ae70528576ef7128acde9f7d9a0b6\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1223645396},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-cli@sha256:69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9\\\",\\\"registry.redhat.io/openshift4/ose-cli@sha256:ef83967297f619f45075e7fd1428a1eb981622a6c174c46fb53b158ed24bed85\\\",\\\"registry.redhat.io/openshift4/ose-cli:latest\\\"],\\\"sizeBytes\\\":584351326},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:43 crc kubenswrapper[4775]: E0321 04:52:43.227988 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:43 crc kubenswrapper[4775]: E0321 04:52:43.228470 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:43 crc kubenswrapper[4775]: E0321 04:52:43.228852 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:43 crc kubenswrapper[4775]: E0321 04:52:43.229213 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:43 crc kubenswrapper[4775]: E0321 04:52:43.229241 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 21 04:52:43 crc kubenswrapper[4775]: I0321 04:52:43.276572 4775 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Mar 21 04:52:43 crc kubenswrapper[4775]: I0321 04:52:43.276665 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Mar 21 04:52:43 crc kubenswrapper[4775]: E0321 04:52:43.340580 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="800ms" Mar 21 04:52:43 crc kubenswrapper[4775]: I0321 04:52:43.592438 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 21 04:52:43 crc kubenswrapper[4775]: I0321 04:52:43.595925 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b2996b891633e8617fdbfa635d270d54374602f00b9f36c117a64b0b6e789a4b"} Mar 21 04:52:43 crc kubenswrapper[4775]: I0321 04:52:43.918065 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 21 04:52:43 crc kubenswrapper[4775]: I0321 04:52:43.919185 4775 status_manager.go:851] "Failed to get status for pod" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:43 crc kubenswrapper[4775]: I0321 04:52:43.919696 4775 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.009603 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc0e122c-1775-4bf1-9025-6288c383b3f2-kubelet-dir\") pod \"bc0e122c-1775-4bf1-9025-6288c383b3f2\" (UID: \"bc0e122c-1775-4bf1-9025-6288c383b3f2\") " Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.010127 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bc0e122c-1775-4bf1-9025-6288c383b3f2-var-lock\") pod \"bc0e122c-1775-4bf1-9025-6288c383b3f2\" (UID: \"bc0e122c-1775-4bf1-9025-6288c383b3f2\") " Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.010176 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc0e122c-1775-4bf1-9025-6288c383b3f2-kube-api-access\") pod \"bc0e122c-1775-4bf1-9025-6288c383b3f2\" (UID: \"bc0e122c-1775-4bf1-9025-6288c383b3f2\") " Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.010624 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc0e122c-1775-4bf1-9025-6288c383b3f2-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bc0e122c-1775-4bf1-9025-6288c383b3f2" (UID: "bc0e122c-1775-4bf1-9025-6288c383b3f2"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.010710 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc0e122c-1775-4bf1-9025-6288c383b3f2-var-lock" (OuterVolumeSpecName: "var-lock") pod "bc0e122c-1775-4bf1-9025-6288c383b3f2" (UID: "bc0e122c-1775-4bf1-9025-6288c383b3f2"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.025568 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc0e122c-1775-4bf1-9025-6288c383b3f2-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bc0e122c-1775-4bf1-9025-6288c383b3f2" (UID: "bc0e122c-1775-4bf1-9025-6288c383b3f2"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.111759 4775 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bc0e122c-1775-4bf1-9025-6288c383b3f2-var-lock\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.111829 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc0e122c-1775-4bf1-9025-6288c383b3f2-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.111843 4775 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc0e122c-1775-4bf1-9025-6288c383b3f2-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:44 crc kubenswrapper[4775]: E0321 04:52:44.141934 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="1.6s" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.279347 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.280140 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.281884 4775 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.282155 4775 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.282357 4775 status_manager.go:851] "Failed to get status for pod" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.313795 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.313861 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.313893 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.314161 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.314190 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.314206 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.415336 4775 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.415379 4775 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.415392 4775 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.541167 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wbtgv" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.541281 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wbtgv" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.582332 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wbtgv" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.582990 4775 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.583643 4775 status_manager.go:851] "Failed to get status for pod" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.583933 4775 status_manager.go:851] "Failed to get status for pod" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" pod="openshift-marketplace/community-operators-wbtgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-wbtgv\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.584264 4775 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.604801 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.605814 4775 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70" exitCode=0 Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.605897 4775 scope.go:117] "RemoveContainer" containerID="aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.606065 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.608929 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.608839 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bc0e122c-1775-4bf1-9025-6288c383b3f2","Type":"ContainerDied","Data":"df0920ca1c2002e3d0e08d9fcffa648209bafd25e83252350dced0f61be9bf09"} Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.609355 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df0920ca1c2002e3d0e08d9fcffa648209bafd25e83252350dced0f61be9bf09" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.620449 4775 status_manager.go:851] "Failed to get status for pod" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.620663 4775 status_manager.go:851] "Failed to get status for pod" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" pod="openshift-marketplace/community-operators-wbtgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-wbtgv\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.620832 4775 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.620980 4775 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.625897 4775 status_manager.go:851] "Failed to get status for pod" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" pod="openshift-marketplace/community-operators-wbtgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-wbtgv\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.626653 4775 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.627076 4775 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.627449 4775 status_manager.go:851] "Failed to get status for pod" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.630341 4775 scope.go:117] "RemoveContainer" containerID="1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.643709 4775 scope.go:117] "RemoveContainer" containerID="fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.659917 4775 scope.go:117] "RemoveContainer" containerID="fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.662718 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wbtgv" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.663577 4775 status_manager.go:851] "Failed to get status for pod" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.663965 4775 status_manager.go:851] "Failed to get status for pod" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" pod="openshift-marketplace/community-operators-wbtgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-wbtgv\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.664473 4775 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.664768 4775 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.686417 4775 scope.go:117] "RemoveContainer" containerID="3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.704468 4775 scope.go:117] "RemoveContainer" containerID="381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.733268 4775 scope.go:117] "RemoveContainer" containerID="aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401" Mar 21 04:52:44 crc kubenswrapper[4775]: E0321 04:52:44.733846 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\": container with ID starting with aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401 not found: ID does not exist" containerID="aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.733888 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401"} err="failed to get container status \"aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\": rpc error: code = NotFound desc = could not find container \"aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401\": container with ID starting with aa32e604ef1a820af34bd67cc5002877686b95aab86676f842ac1761e82cc401 not found: ID does not exist" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.733915 4775 scope.go:117] "RemoveContainer" containerID="1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948" Mar 21 04:52:44 crc kubenswrapper[4775]: E0321 04:52:44.734414 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\": container with ID starting with 1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948 not found: ID does not exist" containerID="1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.734443 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948"} err="failed to get container status \"1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\": rpc error: code = NotFound desc = could not find container \"1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948\": container with ID starting with 1e58586e2fb6aa3271522f23c55846e0e6001fdd90935c516a9a6e7a3e959948 not found: ID does not exist" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.734463 4775 scope.go:117] "RemoveContainer" containerID="fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb" Mar 21 04:52:44 crc kubenswrapper[4775]: E0321 04:52:44.734772 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\": container with ID starting with fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb not found: ID does not exist" containerID="fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.734819 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb"} err="failed to get container status \"fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\": rpc error: code = NotFound desc = could not find container \"fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb\": container with ID starting with fa001dc47bf149b06b215b924a37cac3440e089b361911df93a4c0143a3f53cb not found: ID does not exist" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.734849 4775 scope.go:117] "RemoveContainer" containerID="fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc" Mar 21 04:52:44 crc kubenswrapper[4775]: E0321 04:52:44.735106 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\": container with ID starting with fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc not found: ID does not exist" containerID="fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.735139 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc"} err="failed to get container status \"fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\": rpc error: code = NotFound desc = could not find container \"fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc\": container with ID starting with fa4697d53ba36739e797dace487196678b2dc17d70e745df89372d46af5657bc not found: ID does not exist" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.735154 4775 scope.go:117] "RemoveContainer" containerID="3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70" Mar 21 04:52:44 crc kubenswrapper[4775]: E0321 04:52:44.736405 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\": container with ID starting with 3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70 not found: ID does not exist" containerID="3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.736425 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70"} err="failed to get container status \"3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\": rpc error: code = NotFound desc = could not find container \"3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70\": container with ID starting with 3e874e78570f0b3a6bd4016ccfa09d4c6b81827dc006f3b17f9c93dc65bcac70 not found: ID does not exist" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.736437 4775 scope.go:117] "RemoveContainer" containerID="381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec" Mar 21 04:52:44 crc kubenswrapper[4775]: E0321 04:52:44.736638 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\": container with ID starting with 381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec not found: ID does not exist" containerID="381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.736671 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec"} err="failed to get container status \"381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\": rpc error: code = NotFound desc = could not find container \"381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec\": container with ID starting with 381f5c10200a4e21bc69a5b9f8eca747d803e177b8497ebe7d82f12c98aea3ec not found: ID does not exist" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.961506 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-svf9w" Mar 21 04:52:44 crc kubenswrapper[4775]: I0321 04:52:44.961569 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-svf9w" Mar 21 04:52:45 crc kubenswrapper[4775]: I0321 04:52:45.000088 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-svf9w" Mar 21 04:52:45 crc kubenswrapper[4775]: I0321 04:52:45.000500 4775 status_manager.go:851] "Failed to get status for pod" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" pod="openshift-marketplace/community-operators-wbtgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-wbtgv\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:45 crc kubenswrapper[4775]: I0321 04:52:45.000767 4775 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:45 crc kubenswrapper[4775]: I0321 04:52:45.001061 4775 status_manager.go:851] "Failed to get status for pod" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" pod="openshift-marketplace/community-operators-svf9w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-svf9w\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:45 crc kubenswrapper[4775]: I0321 04:52:45.001444 4775 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:45 crc kubenswrapper[4775]: I0321 04:52:45.001686 4775 status_manager.go:851] "Failed to get status for pod" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:45 crc kubenswrapper[4775]: I0321 04:52:45.660183 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-svf9w" Mar 21 04:52:45 crc kubenswrapper[4775]: I0321 04:52:45.661152 4775 status_manager.go:851] "Failed to get status for pod" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" pod="openshift-marketplace/community-operators-svf9w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-svf9w\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:45 crc kubenswrapper[4775]: I0321 04:52:45.661691 4775 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:45 crc kubenswrapper[4775]: I0321 04:52:45.661952 4775 status_manager.go:851] "Failed to get status for pod" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:45 crc kubenswrapper[4775]: I0321 04:52:45.662416 4775 status_manager.go:851] "Failed to get status for pod" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" pod="openshift-marketplace/community-operators-wbtgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-wbtgv\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:45 crc kubenswrapper[4775]: I0321 04:52:45.662867 4775 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:45 crc kubenswrapper[4775]: I0321 04:52:45.667726 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 21 04:52:45 crc kubenswrapper[4775]: E0321 04:52:45.744252 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="3.2s" Mar 21 04:52:46 crc kubenswrapper[4775]: I0321 04:52:46.722588 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hldm7" Mar 21 04:52:46 crc kubenswrapper[4775]: I0321 04:52:46.723163 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hldm7" Mar 21 04:52:46 crc kubenswrapper[4775]: I0321 04:52:46.781914 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hldm7" Mar 21 04:52:46 crc kubenswrapper[4775]: I0321 04:52:46.782735 4775 status_manager.go:851] "Failed to get status for pod" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" pod="openshift-marketplace/community-operators-wbtgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-wbtgv\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:46 crc kubenswrapper[4775]: I0321 04:52:46.783546 4775 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:46 crc kubenswrapper[4775]: I0321 04:52:46.783864 4775 status_manager.go:851] "Failed to get status for pod" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" pod="openshift-marketplace/community-operators-svf9w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-svf9w\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:46 crc kubenswrapper[4775]: I0321 04:52:46.784192 4775 status_manager.go:851] "Failed to get status for pod" podUID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" pod="openshift-marketplace/redhat-marketplace-hldm7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hldm7\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:46 crc kubenswrapper[4775]: I0321 04:52:46.784527 4775 status_manager.go:851] "Failed to get status for pod" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.165331 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7k9df" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.165377 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7k9df" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.204024 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7k9df" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.204769 4775 status_manager.go:851] "Failed to get status for pod" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.205030 4775 status_manager.go:851] "Failed to get status for pod" podUID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" pod="openshift-marketplace/redhat-marketplace-7k9df" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-7k9df\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.205411 4775 status_manager.go:851] "Failed to get status for pod" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" pod="openshift-marketplace/community-operators-wbtgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-wbtgv\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.205870 4775 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.206090 4775 status_manager.go:851] "Failed to get status for pod" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" pod="openshift-marketplace/community-operators-svf9w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-svf9w\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.206339 4775 status_manager.go:851] "Failed to get status for pod" podUID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" pod="openshift-marketplace/redhat-marketplace-hldm7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hldm7\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: E0321 04:52:47.497481 4775 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.200:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189ec223cf4ca9d3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 04:52:42.288105939 +0000 UTC m=+315.264569563,LastTimestamp:2026-03-21 04:52:42.288105939 +0000 UTC m=+315.264569563,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.540880 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wjfwd" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.540953 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wjfwd" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.592182 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wjfwd" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.592909 4775 status_manager.go:851] "Failed to get status for pod" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" pod="openshift-marketplace/community-operators-wbtgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-wbtgv\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.593446 4775 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.593798 4775 status_manager.go:851] "Failed to get status for pod" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" pod="openshift-marketplace/community-operators-svf9w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-svf9w\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.594450 4775 status_manager.go:851] "Failed to get status for pod" podUID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" pod="openshift-marketplace/redhat-marketplace-hldm7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hldm7\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.594970 4775 status_manager.go:851] "Failed to get status for pod" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.595756 4775 status_manager.go:851] "Failed to get status for pod" podUID="70ad413a-5f81-4094-b2d8-9b89698c6e32" pod="openshift-marketplace/redhat-operators-wjfwd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wjfwd\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.596244 4775 status_manager.go:851] "Failed to get status for pod" podUID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" pod="openshift-marketplace/redhat-marketplace-7k9df" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-7k9df\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.665778 4775 status_manager.go:851] "Failed to get status for pod" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" pod="openshift-marketplace/community-operators-wbtgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-wbtgv\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.666391 4775 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.666755 4775 status_manager.go:851] "Failed to get status for pod" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" pod="openshift-marketplace/community-operators-svf9w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-svf9w\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.668235 4775 status_manager.go:851] "Failed to get status for pod" podUID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" pod="openshift-marketplace/redhat-marketplace-hldm7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hldm7\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.668672 4775 status_manager.go:851] "Failed to get status for pod" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.669099 4775 status_manager.go:851] "Failed to get status for pod" podUID="70ad413a-5f81-4094-b2d8-9b89698c6e32" pod="openshift-marketplace/redhat-operators-wjfwd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wjfwd\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.669902 4775 status_manager.go:851] "Failed to get status for pod" podUID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" pod="openshift-marketplace/redhat-marketplace-7k9df" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-7k9df\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.678540 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wjfwd" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.679214 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hldm7" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.679333 4775 status_manager.go:851] "Failed to get status for pod" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" pod="openshift-marketplace/community-operators-svf9w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-svf9w\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.679942 4775 status_manager.go:851] "Failed to get status for pod" podUID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" pod="openshift-marketplace/redhat-marketplace-hldm7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hldm7\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.680470 4775 status_manager.go:851] "Failed to get status for pod" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.681229 4775 status_manager.go:851] "Failed to get status for pod" podUID="70ad413a-5f81-4094-b2d8-9b89698c6e32" pod="openshift-marketplace/redhat-operators-wjfwd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wjfwd\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.681657 4775 status_manager.go:851] "Failed to get status for pod" podUID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" pod="openshift-marketplace/redhat-marketplace-7k9df" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-7k9df\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.682078 4775 status_manager.go:851] "Failed to get status for pod" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" pod="openshift-marketplace/community-operators-wbtgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-wbtgv\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.682526 4775 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.683321 4775 status_manager.go:851] "Failed to get status for pod" podUID="70ad413a-5f81-4094-b2d8-9b89698c6e32" pod="openshift-marketplace/redhat-operators-wjfwd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wjfwd\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.683906 4775 status_manager.go:851] "Failed to get status for pod" podUID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" pod="openshift-marketplace/redhat-marketplace-7k9df" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-7k9df\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.684377 4775 status_manager.go:851] "Failed to get status for pod" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" pod="openshift-marketplace/community-operators-wbtgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-wbtgv\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.684823 4775 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.685342 4775 status_manager.go:851] "Failed to get status for pod" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" pod="openshift-marketplace/community-operators-svf9w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-svf9w\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.685796 4775 status_manager.go:851] "Failed to get status for pod" podUID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" pod="openshift-marketplace/redhat-marketplace-hldm7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hldm7\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.686084 4775 status_manager.go:851] "Failed to get status for pod" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.688984 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7k9df" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.689345 4775 status_manager.go:851] "Failed to get status for pod" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.689667 4775 status_manager.go:851] "Failed to get status for pod" podUID="70ad413a-5f81-4094-b2d8-9b89698c6e32" pod="openshift-marketplace/redhat-operators-wjfwd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wjfwd\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.690064 4775 status_manager.go:851] "Failed to get status for pod" podUID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" pod="openshift-marketplace/redhat-marketplace-7k9df" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-7k9df\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.690535 4775 status_manager.go:851] "Failed to get status for pod" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" pod="openshift-marketplace/community-operators-wbtgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-wbtgv\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.690815 4775 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.691199 4775 status_manager.go:851] "Failed to get status for pod" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" pod="openshift-marketplace/community-operators-svf9w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-svf9w\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:47 crc kubenswrapper[4775]: I0321 04:52:47.691620 4775 status_manager.go:851] "Failed to get status for pod" podUID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" pod="openshift-marketplace/redhat-marketplace-hldm7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hldm7\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:48 crc kubenswrapper[4775]: E0321 04:52:48.945412 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="6.4s" Mar 21 04:52:53 crc kubenswrapper[4775]: E0321 04:52:53.239429 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:52:53Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:52:53Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:52:53Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T04:52:53Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4869c69128f74d9c3b178ea6c8c8d38df169b6bce05eb821a65f0aaf514c563a\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:f3c2ad90e251062165f8d6623ca4994c0b3e28324e4b5b17fd588b162ec97766\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1746912226},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:04dd6a6688563ae1dbeb924643a49d81150c900e2f7e81aaab8a3586e6f53044\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:bc9a28a1c3a5f16a1873e4919f236a9824c62c6394093adb4850b30daf95761c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1252939492},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:898c67bf7fc973e99114f3148976a6c21ae0dbe413051415588fa9b995f5b331\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:a641939d2096609a4cf6eec872a1476b7c671bfd81cffc2edeb6e9f13c9deeba\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1231028434},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0634296304362d86f6b8d0de6c3e1b327670f86a3e0a664a7fd2fa3984fbbb18\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:1a955538926dc95f0f30489b33552cf1f18ae70528576ef7128acde9f7d9a0b6\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1223645396},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-cli@sha256:69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9\\\",\\\"registry.redhat.io/openshift4/ose-cli@sha256:ef83967297f619f45075e7fd1428a1eb981622a6c174c46fb53b158ed24bed85\\\",\\\"registry.redhat.io/openshift4/ose-cli:latest\\\"],\\\"sizeBytes\\\":584351326},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:53 crc kubenswrapper[4775]: E0321 04:52:53.241214 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:53 crc kubenswrapper[4775]: E0321 04:52:53.241659 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:53 crc kubenswrapper[4775]: E0321 04:52:53.243317 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:53 crc kubenswrapper[4775]: E0321 04:52:53.243546 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:53 crc kubenswrapper[4775]: E0321 04:52:53.243564 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 21 04:52:53 crc kubenswrapper[4775]: I0321 04:52:53.660762 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:52:53 crc kubenswrapper[4775]: I0321 04:52:53.661528 4775 status_manager.go:851] "Failed to get status for pod" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" pod="openshift-marketplace/community-operators-svf9w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-svf9w\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:53 crc kubenswrapper[4775]: I0321 04:52:53.661884 4775 status_manager.go:851] "Failed to get status for pod" podUID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" pod="openshift-marketplace/redhat-marketplace-hldm7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hldm7\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:53 crc kubenswrapper[4775]: I0321 04:52:53.662210 4775 status_manager.go:851] "Failed to get status for pod" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:53 crc kubenswrapper[4775]: I0321 04:52:53.662555 4775 status_manager.go:851] "Failed to get status for pod" podUID="70ad413a-5f81-4094-b2d8-9b89698c6e32" pod="openshift-marketplace/redhat-operators-wjfwd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wjfwd\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:53 crc kubenswrapper[4775]: I0321 04:52:53.662918 4775 status_manager.go:851] "Failed to get status for pod" podUID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" pod="openshift-marketplace/redhat-marketplace-7k9df" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-7k9df\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:53 crc kubenswrapper[4775]: I0321 04:52:53.663193 4775 status_manager.go:851] "Failed to get status for pod" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" pod="openshift-marketplace/community-operators-wbtgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-wbtgv\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:53 crc kubenswrapper[4775]: I0321 04:52:53.663581 4775 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:53 crc kubenswrapper[4775]: I0321 04:52:53.680112 4775 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cbc8474b-4360-449d-ab37-ba14ca1ac5ed" Mar 21 04:52:53 crc kubenswrapper[4775]: I0321 04:52:53.680178 4775 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cbc8474b-4360-449d-ab37-ba14ca1ac5ed" Mar 21 04:52:53 crc kubenswrapper[4775]: E0321 04:52:53.680550 4775 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:52:53 crc kubenswrapper[4775]: I0321 04:52:53.680988 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:52:53 crc kubenswrapper[4775]: W0321 04:52:53.699148 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-1f41636626addbc72b3497e2fe8340a674aebde57e937a756d885b1e10949d0b WatchSource:0}: Error finding container 1f41636626addbc72b3497e2fe8340a674aebde57e937a756d885b1e10949d0b: Status 404 returned error can't find the container with id 1f41636626addbc72b3497e2fe8340a674aebde57e937a756d885b1e10949d0b Mar 21 04:52:54 crc kubenswrapper[4775]: I0321 04:52:54.668879 4775 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="378b58294d33a1a18c3716c9a58fc22e327620d44b86d33adc8bb7640257fe85" exitCode=0 Mar 21 04:52:54 crc kubenswrapper[4775]: I0321 04:52:54.668979 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"378b58294d33a1a18c3716c9a58fc22e327620d44b86d33adc8bb7640257fe85"} Mar 21 04:52:54 crc kubenswrapper[4775]: I0321 04:52:54.669243 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1f41636626addbc72b3497e2fe8340a674aebde57e937a756d885b1e10949d0b"} Mar 21 04:52:54 crc kubenswrapper[4775]: I0321 04:52:54.669513 4775 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cbc8474b-4360-449d-ab37-ba14ca1ac5ed" Mar 21 04:52:54 crc kubenswrapper[4775]: I0321 04:52:54.669529 4775 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cbc8474b-4360-449d-ab37-ba14ca1ac5ed" Mar 21 04:52:54 crc kubenswrapper[4775]: E0321 04:52:54.669987 4775 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:52:54 crc kubenswrapper[4775]: I0321 04:52:54.671754 4775 status_manager.go:851] "Failed to get status for pod" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" pod="openshift-marketplace/community-operators-wbtgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-wbtgv\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:54 crc kubenswrapper[4775]: I0321 04:52:54.672261 4775 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:54 crc kubenswrapper[4775]: I0321 04:52:54.672576 4775 status_manager.go:851] "Failed to get status for pod" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" pod="openshift-marketplace/community-operators-svf9w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-svf9w\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:54 crc kubenswrapper[4775]: I0321 04:52:54.672837 4775 status_manager.go:851] "Failed to get status for pod" podUID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" pod="openshift-marketplace/redhat-marketplace-hldm7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hldm7\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:54 crc kubenswrapper[4775]: I0321 04:52:54.673051 4775 status_manager.go:851] "Failed to get status for pod" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:54 crc kubenswrapper[4775]: I0321 04:52:54.673335 4775 status_manager.go:851] "Failed to get status for pod" podUID="70ad413a-5f81-4094-b2d8-9b89698c6e32" pod="openshift-marketplace/redhat-operators-wjfwd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wjfwd\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:54 crc kubenswrapper[4775]: I0321 04:52:54.673813 4775 status_manager.go:851] "Failed to get status for pod" podUID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" pod="openshift-marketplace/redhat-marketplace-7k9df" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-7k9df\": dial tcp 38.102.83.200:6443: connect: connection refused" Mar 21 04:52:55 crc kubenswrapper[4775]: I0321 04:52:55.677992 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1190cd41522dfa2bcdaf0b043b1b6e08f1d536aefb6a5b94f237c6bf9a884f03"} Mar 21 04:52:55 crc kubenswrapper[4775]: I0321 04:52:55.678603 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c4b28862738ff9483688e677523118653d5826d836f2a12d150db9feaea262ba"} Mar 21 04:52:55 crc kubenswrapper[4775]: I0321 04:52:55.678618 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"acf01814aa883f6c53895f25cdd6f332afbc3dce1f7d957c3686d6cfbe640bc7"} Mar 21 04:52:55 crc kubenswrapper[4775]: I0321 04:52:55.678648 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9bc226a97f0f60ca373e5050ef0d1e1c7a82c039c94c9976d98d9b115d6d2dc9"} Mar 21 04:52:56 crc kubenswrapper[4775]: I0321 04:52:56.689948 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 21 04:52:56 crc kubenswrapper[4775]: I0321 04:52:56.691199 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 21 04:52:56 crc kubenswrapper[4775]: I0321 04:52:56.691285 4775 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="75cf7f90f0971bce2e51690954080e297523b92a5588c8effa75832879f0dd6b" exitCode=1 Mar 21 04:52:56 crc kubenswrapper[4775]: I0321 04:52:56.691385 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"75cf7f90f0971bce2e51690954080e297523b92a5588c8effa75832879f0dd6b"} Mar 21 04:52:56 crc kubenswrapper[4775]: I0321 04:52:56.692902 4775 scope.go:117] "RemoveContainer" containerID="75cf7f90f0971bce2e51690954080e297523b92a5588c8effa75832879f0dd6b" Mar 21 04:52:56 crc kubenswrapper[4775]: I0321 04:52:56.697705 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"34b753c2ca5cb3f8746f3ffe7cc0f8289b2f44cd7679e8ac546b87a7bd71bf17"} Mar 21 04:52:56 crc kubenswrapper[4775]: I0321 04:52:56.698034 4775 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cbc8474b-4360-449d-ab37-ba14ca1ac5ed" Mar 21 04:52:56 crc kubenswrapper[4775]: I0321 04:52:56.698068 4775 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cbc8474b-4360-449d-ab37-ba14ca1ac5ed" Mar 21 04:52:56 crc kubenswrapper[4775]: I0321 04:52:56.698338 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:52:57 crc kubenswrapper[4775]: I0321 04:52:57.705548 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 21 04:52:57 crc kubenswrapper[4775]: I0321 04:52:57.706437 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 21 04:52:57 crc kubenswrapper[4775]: I0321 04:52:57.706487 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d8a313e0d3f3cfd85aab72e5d3551cfa434d04205dc6565ceaaff0c072352a77"} Mar 21 04:52:58 crc kubenswrapper[4775]: I0321 04:52:58.681881 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:52:58 crc kubenswrapper[4775]: I0321 04:52:58.681964 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:52:58 crc kubenswrapper[4775]: I0321 04:52:58.688945 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:53:01 crc kubenswrapper[4775]: I0321 04:53:01.709157 4775 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:53:01 crc kubenswrapper[4775]: I0321 04:53:01.730532 4775 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cbc8474b-4360-449d-ab37-ba14ca1ac5ed" Mar 21 04:53:01 crc kubenswrapper[4775]: I0321 04:53:01.730578 4775 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cbc8474b-4360-449d-ab37-ba14ca1ac5ed" Mar 21 04:53:01 crc kubenswrapper[4775]: I0321 04:53:01.737545 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:53:01 crc kubenswrapper[4775]: I0321 04:53:01.905094 4775 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="88571c4e-5944-4868-b298-4cacd3831adc" Mar 21 04:53:02 crc kubenswrapper[4775]: I0321 04:53:02.467052 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:53:02 crc kubenswrapper[4775]: I0321 04:53:02.467227 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:53:02 crc kubenswrapper[4775]: I0321 04:53:02.469340 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 21 04:53:02 crc kubenswrapper[4775]: I0321 04:53:02.469368 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 21 04:53:02 crc kubenswrapper[4775]: I0321 04:53:02.479749 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:53:02 crc kubenswrapper[4775]: I0321 04:53:02.485270 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:53:02 crc kubenswrapper[4775]: I0321 04:53:02.736491 4775 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cbc8474b-4360-449d-ab37-ba14ca1ac5ed" Mar 21 04:53:02 crc kubenswrapper[4775]: I0321 04:53:02.737248 4775 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cbc8474b-4360-449d-ab37-ba14ca1ac5ed" Mar 21 04:53:02 crc kubenswrapper[4775]: I0321 04:53:02.739738 4775 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="88571c4e-5944-4868-b298-4cacd3831adc" Mar 21 04:53:02 crc kubenswrapper[4775]: I0321 04:53:02.783736 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 04:53:03 crc kubenswrapper[4775]: W0321 04:53:03.176623 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-4c2f493ebe9ff005b4a2558285dab577d4b15c2f387c7d7ba5d4f6b52621d799 WatchSource:0}: Error finding container 4c2f493ebe9ff005b4a2558285dab577d4b15c2f387c7d7ba5d4f6b52621d799: Status 404 returned error can't find the container with id 4c2f493ebe9ff005b4a2558285dab577d4b15c2f387c7d7ba5d4f6b52621d799 Mar 21 04:53:03 crc kubenswrapper[4775]: I0321 04:53:03.744060 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"aab1d0277ec455f9d7273f185cd4c536fde8444247b04b3d86da1923ad2d0d0e"} Mar 21 04:53:03 crc kubenswrapper[4775]: I0321 04:53:03.744463 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"4c2f493ebe9ff005b4a2558285dab577d4b15c2f387c7d7ba5d4f6b52621d799"} Mar 21 04:53:05 crc kubenswrapper[4775]: I0321 04:53:05.017236 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:53:05 crc kubenswrapper[4775]: I0321 04:53:05.616337 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:53:05 crc kubenswrapper[4775]: I0321 04:53:05.620063 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:53:06 crc kubenswrapper[4775]: I0321 04:53:06.768647 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 04:53:11 crc kubenswrapper[4775]: I0321 04:53:11.010301 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 21 04:53:11 crc kubenswrapper[4775]: I0321 04:53:11.734496 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 21 04:53:12 crc kubenswrapper[4775]: I0321 04:53:12.552773 4775 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 21 04:53:12 crc kubenswrapper[4775]: I0321 04:53:12.791462 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 21 04:53:12 crc kubenswrapper[4775]: I0321 04:53:12.871583 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 21 04:53:12 crc kubenswrapper[4775]: I0321 04:53:12.989080 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 21 04:53:13 crc kubenswrapper[4775]: I0321 04:53:13.091925 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 21 04:53:13 crc kubenswrapper[4775]: I0321 04:53:13.176492 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 21 04:53:13 crc kubenswrapper[4775]: I0321 04:53:13.319205 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 21 04:53:13 crc kubenswrapper[4775]: I0321 04:53:13.319783 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 21 04:53:13 crc kubenswrapper[4775]: I0321 04:53:13.467627 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 21 04:53:14 crc kubenswrapper[4775]: I0321 04:53:14.012458 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 21 04:53:14 crc kubenswrapper[4775]: I0321 04:53:14.335931 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 21 04:53:14 crc kubenswrapper[4775]: I0321 04:53:14.461972 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 21 04:53:14 crc kubenswrapper[4775]: I0321 04:53:14.481924 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 21 04:53:14 crc kubenswrapper[4775]: I0321 04:53:14.618231 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 21 04:53:14 crc kubenswrapper[4775]: I0321 04:53:14.762179 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 21 04:53:14 crc kubenswrapper[4775]: I0321 04:53:14.773486 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 21 04:53:14 crc kubenswrapper[4775]: I0321 04:53:14.833888 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 21 04:53:14 crc kubenswrapper[4775]: I0321 04:53:14.959497 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 21 04:53:15 crc kubenswrapper[4775]: I0321 04:53:15.292571 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 21 04:53:15 crc kubenswrapper[4775]: I0321 04:53:15.311453 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 21 04:53:15 crc kubenswrapper[4775]: I0321 04:53:15.453755 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 21 04:53:15 crc kubenswrapper[4775]: I0321 04:53:15.628622 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 21 04:53:15 crc kubenswrapper[4775]: I0321 04:53:15.726398 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 21 04:53:15 crc kubenswrapper[4775]: I0321 04:53:15.765665 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 21 04:53:15 crc kubenswrapper[4775]: I0321 04:53:15.879044 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 21 04:53:15 crc kubenswrapper[4775]: I0321 04:53:15.906721 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 21 04:53:15 crc kubenswrapper[4775]: I0321 04:53:15.911395 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 21 04:53:16 crc kubenswrapper[4775]: I0321 04:53:16.004179 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 21 04:53:16 crc kubenswrapper[4775]: I0321 04:53:16.070858 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 21 04:53:16 crc kubenswrapper[4775]: I0321 04:53:16.121835 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 21 04:53:16 crc kubenswrapper[4775]: I0321 04:53:16.130563 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 21 04:53:16 crc kubenswrapper[4775]: I0321 04:53:16.173724 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 21 04:53:16 crc kubenswrapper[4775]: I0321 04:53:16.197087 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 21 04:53:16 crc kubenswrapper[4775]: I0321 04:53:16.218498 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 21 04:53:16 crc kubenswrapper[4775]: I0321 04:53:16.425797 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 21 04:53:16 crc kubenswrapper[4775]: I0321 04:53:16.554981 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 21 04:53:16 crc kubenswrapper[4775]: I0321 04:53:16.768911 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 21 04:53:16 crc kubenswrapper[4775]: I0321 04:53:16.793529 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 21 04:53:16 crc kubenswrapper[4775]: I0321 04:53:16.981987 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 21 04:53:16 crc kubenswrapper[4775]: I0321 04:53:16.997942 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 21 04:53:17 crc kubenswrapper[4775]: I0321 04:53:17.112261 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 21 04:53:17 crc kubenswrapper[4775]: I0321 04:53:17.138728 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 21 04:53:17 crc kubenswrapper[4775]: I0321 04:53:17.541150 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 21 04:53:17 crc kubenswrapper[4775]: I0321 04:53:17.599222 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 21 04:53:17 crc kubenswrapper[4775]: I0321 04:53:17.784345 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 21 04:53:17 crc kubenswrapper[4775]: I0321 04:53:17.797922 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 21 04:53:17 crc kubenswrapper[4775]: I0321 04:53:17.877517 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 21 04:53:17 crc kubenswrapper[4775]: I0321 04:53:17.961544 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 21 04:53:18 crc kubenswrapper[4775]: I0321 04:53:18.030394 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 21 04:53:18 crc kubenswrapper[4775]: I0321 04:53:18.035553 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 21 04:53:18 crc kubenswrapper[4775]: I0321 04:53:18.139015 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 21 04:53:18 crc kubenswrapper[4775]: I0321 04:53:18.156646 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 21 04:53:18 crc kubenswrapper[4775]: I0321 04:53:18.272332 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 21 04:53:18 crc kubenswrapper[4775]: I0321 04:53:18.373986 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 21 04:53:18 crc kubenswrapper[4775]: I0321 04:53:18.385698 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 21 04:53:18 crc kubenswrapper[4775]: I0321 04:53:18.452102 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 21 04:53:18 crc kubenswrapper[4775]: I0321 04:53:18.521654 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 21 04:53:18 crc kubenswrapper[4775]: I0321 04:53:18.525093 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 21 04:53:18 crc kubenswrapper[4775]: I0321 04:53:18.671022 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 21 04:53:18 crc kubenswrapper[4775]: I0321 04:53:18.736086 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 21 04:53:18 crc kubenswrapper[4775]: I0321 04:53:18.742490 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 21 04:53:18 crc kubenswrapper[4775]: I0321 04:53:18.771879 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 21 04:53:18 crc kubenswrapper[4775]: I0321 04:53:18.775516 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 21 04:53:18 crc kubenswrapper[4775]: I0321 04:53:18.783817 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 21 04:53:18 crc kubenswrapper[4775]: I0321 04:53:18.902269 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 21 04:53:18 crc kubenswrapper[4775]: I0321 04:53:18.903755 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 21 04:53:18 crc kubenswrapper[4775]: I0321 04:53:18.904875 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 21 04:53:19 crc kubenswrapper[4775]: I0321 04:53:19.017615 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 21 04:53:19 crc kubenswrapper[4775]: I0321 04:53:19.111882 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 21 04:53:19 crc kubenswrapper[4775]: I0321 04:53:19.351525 4775 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 21 04:53:19 crc kubenswrapper[4775]: I0321 04:53:19.458622 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 21 04:53:19 crc kubenswrapper[4775]: I0321 04:53:19.486107 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 21 04:53:19 crc kubenswrapper[4775]: I0321 04:53:19.509235 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 21 04:53:19 crc kubenswrapper[4775]: I0321 04:53:19.509902 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 21 04:53:19 crc kubenswrapper[4775]: I0321 04:53:19.530012 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 21 04:53:19 crc kubenswrapper[4775]: I0321 04:53:19.590897 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 21 04:53:19 crc kubenswrapper[4775]: I0321 04:53:19.647411 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 21 04:53:19 crc kubenswrapper[4775]: I0321 04:53:19.649762 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 21 04:53:19 crc kubenswrapper[4775]: I0321 04:53:19.695213 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 21 04:53:19 crc kubenswrapper[4775]: I0321 04:53:19.767496 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 21 04:53:19 crc kubenswrapper[4775]: I0321 04:53:19.778594 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 21 04:53:19 crc kubenswrapper[4775]: I0321 04:53:19.786929 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 21 04:53:19 crc kubenswrapper[4775]: I0321 04:53:19.844694 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 21 04:53:19 crc kubenswrapper[4775]: I0321 04:53:19.868622 4775 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 21 04:53:19 crc kubenswrapper[4775]: I0321 04:53:19.886687 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 21 04:53:19 crc kubenswrapper[4775]: I0321 04:53:19.888034 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 21 04:53:19 crc kubenswrapper[4775]: I0321 04:53:19.977204 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 21 04:53:20 crc kubenswrapper[4775]: I0321 04:53:20.081426 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 21 04:53:20 crc kubenswrapper[4775]: I0321 04:53:20.088101 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 21 04:53:20 crc kubenswrapper[4775]: I0321 04:53:20.090860 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 21 04:53:20 crc kubenswrapper[4775]: I0321 04:53:20.092212 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 21 04:53:20 crc kubenswrapper[4775]: I0321 04:53:20.138748 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 21 04:53:20 crc kubenswrapper[4775]: I0321 04:53:20.176327 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 21 04:53:20 crc kubenswrapper[4775]: I0321 04:53:20.178952 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 21 04:53:20 crc kubenswrapper[4775]: I0321 04:53:20.276585 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 21 04:53:20 crc kubenswrapper[4775]: I0321 04:53:20.277741 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 21 04:53:20 crc kubenswrapper[4775]: I0321 04:53:20.290320 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 21 04:53:20 crc kubenswrapper[4775]: I0321 04:53:20.315507 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 21 04:53:20 crc kubenswrapper[4775]: I0321 04:53:20.644692 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 21 04:53:20 crc kubenswrapper[4775]: I0321 04:53:20.658918 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 21 04:53:20 crc kubenswrapper[4775]: I0321 04:53:20.663401 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 21 04:53:20 crc kubenswrapper[4775]: I0321 04:53:20.828439 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 21 04:53:20 crc kubenswrapper[4775]: I0321 04:53:20.841758 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 21 04:53:20 crc kubenswrapper[4775]: I0321 04:53:20.844411 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 21 04:53:20 crc kubenswrapper[4775]: I0321 04:53:20.994260 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.021807 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.078845 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.252477 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.278648 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.286965 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.338194 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.380836 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.461840 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.491245 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.517088 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.671815 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.707550 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.741701 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.749140 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.801670 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.845776 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.875545 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.929268 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.941343 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.941448 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 21 04:53:21 crc kubenswrapper[4775]: I0321 04:53:21.990446 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 21 04:53:22 crc kubenswrapper[4775]: I0321 04:53:22.048895 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 21 04:53:22 crc kubenswrapper[4775]: I0321 04:53:22.201615 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 21 04:53:22 crc kubenswrapper[4775]: I0321 04:53:22.225782 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 21 04:53:22 crc kubenswrapper[4775]: I0321 04:53:22.260249 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 21 04:53:22 crc kubenswrapper[4775]: I0321 04:53:22.303467 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 21 04:53:22 crc kubenswrapper[4775]: I0321 04:53:22.319863 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 21 04:53:22 crc kubenswrapper[4775]: I0321 04:53:22.379488 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 21 04:53:22 crc kubenswrapper[4775]: I0321 04:53:22.417876 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 21 04:53:22 crc kubenswrapper[4775]: I0321 04:53:22.450330 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 21 04:53:22 crc kubenswrapper[4775]: I0321 04:53:22.470040 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 21 04:53:22 crc kubenswrapper[4775]: I0321 04:53:22.527035 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 21 04:53:22 crc kubenswrapper[4775]: I0321 04:53:22.644640 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 21 04:53:22 crc kubenswrapper[4775]: I0321 04:53:22.650309 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 21 04:53:22 crc kubenswrapper[4775]: I0321 04:53:22.692262 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 21 04:53:22 crc kubenswrapper[4775]: I0321 04:53:22.779507 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 21 04:53:22 crc kubenswrapper[4775]: I0321 04:53:22.819401 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 21 04:53:22 crc kubenswrapper[4775]: I0321 04:53:22.924697 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 21 04:53:22 crc kubenswrapper[4775]: I0321 04:53:22.937997 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 21 04:53:22 crc kubenswrapper[4775]: I0321 04:53:22.963812 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 21 04:53:22 crc kubenswrapper[4775]: I0321 04:53:22.979770 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.082463 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.083031 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.223453 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.246683 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.375740 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.398759 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.535494 4775 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.536022 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=42.536008707 podStartE2EDuration="42.536008707s" podCreationTimestamp="2026-03-21 04:52:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:53:01.759528412 +0000 UTC m=+334.735992046" watchObservedRunningTime="2026-03-21 04:53:23.536008707 +0000 UTC m=+356.512472331" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.539310 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.539354 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7d8984bd85-5vxq9","openshift-kube-apiserver/kube-apiserver-crc"] Mar 21 04:53:23 crc kubenswrapper[4775]: E0321 04:53:23.539518 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" containerName="installer" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.539533 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" containerName="installer" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.539640 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc0e122c-1775-4bf1-9025-6288c383b3f2" containerName="installer" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.539984 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.539982 4775 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cbc8474b-4360-449d-ab37-ba14ca1ac5ed" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.540372 4775 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cbc8474b-4360-449d-ab37-ba14ca1ac5ed" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.543027 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.543296 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.543349 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.544522 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.544542 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.544619 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.544902 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.546307 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.546381 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.546436 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.546542 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.546709 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.547195 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.552464 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.558098 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.561107 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.561308 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.568941 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=22.568924427 podStartE2EDuration="22.568924427s" podCreationTimestamp="2026-03-21 04:53:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:53:23.567080794 +0000 UTC m=+356.543544428" watchObservedRunningTime="2026-03-21 04:53:23.568924427 +0000 UTC m=+356.545388051" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.588604 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.641683 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.641759 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-audit-dir\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.641848 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-session\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.641949 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.641973 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.641993 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.642010 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.642043 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-audit-policies\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.642064 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.642158 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2cqt\" (UniqueName: \"kubernetes.io/projected/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-kube-api-access-k2cqt\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.642210 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.642247 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-user-template-login\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.642266 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-user-template-error\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.642294 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.648498 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.681281 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.734281 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.743671 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-audit-dir\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.743708 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-session\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.743747 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.743765 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.743785 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.743807 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.743828 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-audit-policies\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.743830 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-audit-dir\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.743845 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.743958 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2cqt\" (UniqueName: \"kubernetes.io/projected/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-kube-api-access-k2cqt\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.743980 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.744001 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-user-template-login\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.744018 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-user-template-error\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.744038 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.744065 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.745812 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.746094 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.746271 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.746508 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-audit-policies\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.751907 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.752849 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-session\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.752865 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.753202 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.753979 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-user-template-login\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.754652 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-user-template-error\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.756852 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.759666 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.763803 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2cqt\" (UniqueName: \"kubernetes.io/projected/de6d7fec-5fbc-4fc4-aae8-ecea20324c19-kube-api-access-k2cqt\") pod \"oauth-openshift-7d8984bd85-5vxq9\" (UID: \"de6d7fec-5fbc-4fc4-aae8-ecea20324c19\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.851679 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.864669 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:23 crc kubenswrapper[4775]: I0321 04:53:23.969221 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.027902 4775 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.028203 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://b2996b891633e8617fdbfa635d270d54374602f00b9f36c117a64b0b6e789a4b" gracePeriod=5 Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.072166 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.110416 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.130194 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.176336 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.188551 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.198770 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.209608 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.217402 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.313839 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.343814 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7d8984bd85-5vxq9"] Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.366547 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.392107 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.406466 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.450837 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.468861 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.613374 4775 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.613525 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.623916 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.676457 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.677379 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.758946 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.868724 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.873331 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" event={"ID":"de6d7fec-5fbc-4fc4-aae8-ecea20324c19","Type":"ContainerStarted","Data":"83bd7fd026ec0926eed8900ca51da6279293b19fe53634179a6f0265deef8faf"} Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.873394 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" event={"ID":"de6d7fec-5fbc-4fc4-aae8-ecea20324c19","Type":"ContainerStarted","Data":"d89bd894b9223c35f5fe3e89582520f5860fac99b7d3bfea7b88fd00a356bb4b"} Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.873762 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.888981 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.894738 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.895956 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" podStartSLOduration=78.895941402 podStartE2EDuration="1m18.895941402s" podCreationTimestamp="2026-03-21 04:52:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:53:24.893210174 +0000 UTC m=+357.869673818" watchObservedRunningTime="2026-03-21 04:53:24.895941402 +0000 UTC m=+357.872405026" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.916741 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.951796 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.956729 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.960748 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 21 04:53:24 crc kubenswrapper[4775]: I0321 04:53:24.993137 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7d8984bd85-5vxq9" Mar 21 04:53:25 crc kubenswrapper[4775]: I0321 04:53:25.089582 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 21 04:53:25 crc kubenswrapper[4775]: I0321 04:53:25.127632 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 21 04:53:25 crc kubenswrapper[4775]: I0321 04:53:25.177262 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 21 04:53:25 crc kubenswrapper[4775]: I0321 04:53:25.183240 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 21 04:53:25 crc kubenswrapper[4775]: I0321 04:53:25.199485 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 21 04:53:25 crc kubenswrapper[4775]: I0321 04:53:25.241369 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 21 04:53:25 crc kubenswrapper[4775]: I0321 04:53:25.291187 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 21 04:53:25 crc kubenswrapper[4775]: I0321 04:53:25.682513 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 21 04:53:25 crc kubenswrapper[4775]: I0321 04:53:25.876371 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 21 04:53:25 crc kubenswrapper[4775]: I0321 04:53:25.884981 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 21 04:53:25 crc kubenswrapper[4775]: I0321 04:53:25.941423 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 21 04:53:26 crc kubenswrapper[4775]: I0321 04:53:26.046002 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 21 04:53:26 crc kubenswrapper[4775]: I0321 04:53:26.102756 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 21 04:53:26 crc kubenswrapper[4775]: I0321 04:53:26.178233 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 21 04:53:26 crc kubenswrapper[4775]: I0321 04:53:26.480355 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 21 04:53:26 crc kubenswrapper[4775]: I0321 04:53:26.488980 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 21 04:53:26 crc kubenswrapper[4775]: I0321 04:53:26.565454 4775 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 21 04:53:26 crc kubenswrapper[4775]: I0321 04:53:26.596310 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 21 04:53:26 crc kubenswrapper[4775]: I0321 04:53:26.629777 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 21 04:53:27 crc kubenswrapper[4775]: I0321 04:53:27.095244 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 21 04:53:27 crc kubenswrapper[4775]: I0321 04:53:27.097572 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 21 04:53:27 crc kubenswrapper[4775]: I0321 04:53:27.151441 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 21 04:53:27 crc kubenswrapper[4775]: I0321 04:53:27.167839 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 21 04:53:27 crc kubenswrapper[4775]: I0321 04:53:27.381788 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 21 04:53:27 crc kubenswrapper[4775]: I0321 04:53:27.477238 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 21 04:53:27 crc kubenswrapper[4775]: I0321 04:53:27.581566 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 21 04:53:27 crc kubenswrapper[4775]: I0321 04:53:27.597900 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 21 04:53:27 crc kubenswrapper[4775]: I0321 04:53:27.687983 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 21 04:53:27 crc kubenswrapper[4775]: I0321 04:53:27.783295 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 21 04:53:27 crc kubenswrapper[4775]: I0321 04:53:27.855096 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 21 04:53:27 crc kubenswrapper[4775]: I0321 04:53:27.962987 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 21 04:53:28 crc kubenswrapper[4775]: I0321 04:53:28.009950 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 21 04:53:28 crc kubenswrapper[4775]: I0321 04:53:28.056556 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 21 04:53:28 crc kubenswrapper[4775]: I0321 04:53:28.353155 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 21 04:53:28 crc kubenswrapper[4775]: I0321 04:53:28.415174 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 21 04:53:28 crc kubenswrapper[4775]: I0321 04:53:28.449163 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 21 04:53:28 crc kubenswrapper[4775]: I0321 04:53:28.629690 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 21 04:53:28 crc kubenswrapper[4775]: I0321 04:53:28.673206 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 21 04:53:28 crc kubenswrapper[4775]: I0321 04:53:28.941053 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.010465 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.127489 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.377691 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.621457 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.621599 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.670916 4775 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.682370 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.682407 4775 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="bd82783d-b559-4327-af5f-eb6d13c77df1" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.687472 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.687530 4775 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="bd82783d-b559-4327-af5f-eb6d13c77df1" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.694482 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.714759 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.729369 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.729524 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.729523 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.729692 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.729581 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.729768 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.729826 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.729921 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.730105 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.730341 4775 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.730387 4775 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.730413 4775 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.730436 4775 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.739676 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.831076 4775 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.905429 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.905523 4775 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="b2996b891633e8617fdbfa635d270d54374602f00b9f36c117a64b0b6e789a4b" exitCode=137 Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.905596 4775 scope.go:117] "RemoveContainer" containerID="b2996b891633e8617fdbfa635d270d54374602f00b9f36c117a64b0b6e789a4b" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.905595 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.932416 4775 scope.go:117] "RemoveContainer" containerID="b2996b891633e8617fdbfa635d270d54374602f00b9f36c117a64b0b6e789a4b" Mar 21 04:53:29 crc kubenswrapper[4775]: E0321 04:53:29.933654 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2996b891633e8617fdbfa635d270d54374602f00b9f36c117a64b0b6e789a4b\": container with ID starting with b2996b891633e8617fdbfa635d270d54374602f00b9f36c117a64b0b6e789a4b not found: ID does not exist" containerID="b2996b891633e8617fdbfa635d270d54374602f00b9f36c117a64b0b6e789a4b" Mar 21 04:53:29 crc kubenswrapper[4775]: I0321 04:53:29.933777 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2996b891633e8617fdbfa635d270d54374602f00b9f36c117a64b0b6e789a4b"} err="failed to get container status \"b2996b891633e8617fdbfa635d270d54374602f00b9f36c117a64b0b6e789a4b\": rpc error: code = NotFound desc = could not find container \"b2996b891633e8617fdbfa635d270d54374602f00b9f36c117a64b0b6e789a4b\": container with ID starting with b2996b891633e8617fdbfa635d270d54374602f00b9f36c117a64b0b6e789a4b not found: ID does not exist" Mar 21 04:53:30 crc kubenswrapper[4775]: I0321 04:53:30.134638 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 21 04:53:30 crc kubenswrapper[4775]: I0321 04:53:30.682047 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 21 04:53:31 crc kubenswrapper[4775]: I0321 04:53:31.668832 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.379850 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6f88f547f-kjxrc"] Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.381216 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" podUID="7da7ce8b-402a-40f7-a9ce-340524ad8573" containerName="controller-manager" containerID="cri-o://b57b29e11c242340693c81affc2d61431477f8ba75c091165a111504297edd4f" gracePeriod=30 Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.487029 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z"] Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.487318 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" podUID="1882b1c5-2465-40df-9306-e5792e0d9f2f" containerName="route-controller-manager" containerID="cri-o://ae81d0a5c5add9cc57df7dd09c08b1e07e8506e8be71c91d0656558f8529bd3b" gracePeriod=30 Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.769546 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.877904 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.903917 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flbhf\" (UniqueName: \"kubernetes.io/projected/7da7ce8b-402a-40f7-a9ce-340524ad8573-kube-api-access-flbhf\") pod \"7da7ce8b-402a-40f7-a9ce-340524ad8573\" (UID: \"7da7ce8b-402a-40f7-a9ce-340524ad8573\") " Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.904018 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da7ce8b-402a-40f7-a9ce-340524ad8573-config\") pod \"7da7ce8b-402a-40f7-a9ce-340524ad8573\" (UID: \"7da7ce8b-402a-40f7-a9ce-340524ad8573\") " Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.904057 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7da7ce8b-402a-40f7-a9ce-340524ad8573-serving-cert\") pod \"7da7ce8b-402a-40f7-a9ce-340524ad8573\" (UID: \"7da7ce8b-402a-40f7-a9ce-340524ad8573\") " Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.904086 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7da7ce8b-402a-40f7-a9ce-340524ad8573-client-ca\") pod \"7da7ce8b-402a-40f7-a9ce-340524ad8573\" (UID: \"7da7ce8b-402a-40f7-a9ce-340524ad8573\") " Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.904157 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7da7ce8b-402a-40f7-a9ce-340524ad8573-proxy-ca-bundles\") pod \"7da7ce8b-402a-40f7-a9ce-340524ad8573\" (UID: \"7da7ce8b-402a-40f7-a9ce-340524ad8573\") " Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.905181 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7da7ce8b-402a-40f7-a9ce-340524ad8573-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7da7ce8b-402a-40f7-a9ce-340524ad8573" (UID: "7da7ce8b-402a-40f7-a9ce-340524ad8573"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.905493 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7da7ce8b-402a-40f7-a9ce-340524ad8573-client-ca" (OuterVolumeSpecName: "client-ca") pod "7da7ce8b-402a-40f7-a9ce-340524ad8573" (UID: "7da7ce8b-402a-40f7-a9ce-340524ad8573"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.905817 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7da7ce8b-402a-40f7-a9ce-340524ad8573-config" (OuterVolumeSpecName: "config") pod "7da7ce8b-402a-40f7-a9ce-340524ad8573" (UID: "7da7ce8b-402a-40f7-a9ce-340524ad8573"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.909645 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7da7ce8b-402a-40f7-a9ce-340524ad8573-kube-api-access-flbhf" (OuterVolumeSpecName: "kube-api-access-flbhf") pod "7da7ce8b-402a-40f7-a9ce-340524ad8573" (UID: "7da7ce8b-402a-40f7-a9ce-340524ad8573"). InnerVolumeSpecName "kube-api-access-flbhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.909923 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7da7ce8b-402a-40f7-a9ce-340524ad8573-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7da7ce8b-402a-40f7-a9ce-340524ad8573" (UID: "7da7ce8b-402a-40f7-a9ce-340524ad8573"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.931344 4775 generic.go:334] "Generic (PLEG): container finished" podID="1882b1c5-2465-40df-9306-e5792e0d9f2f" containerID="ae81d0a5c5add9cc57df7dd09c08b1e07e8506e8be71c91d0656558f8529bd3b" exitCode=0 Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.931422 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.931433 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" event={"ID":"1882b1c5-2465-40df-9306-e5792e0d9f2f","Type":"ContainerDied","Data":"ae81d0a5c5add9cc57df7dd09c08b1e07e8506e8be71c91d0656558f8529bd3b"} Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.931534 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z" event={"ID":"1882b1c5-2465-40df-9306-e5792e0d9f2f","Type":"ContainerDied","Data":"1a6793f0695cc3d26a11dee78d3b1e2f638d464d822526dcc3f7aa8f363599f1"} Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.931553 4775 scope.go:117] "RemoveContainer" containerID="ae81d0a5c5add9cc57df7dd09c08b1e07e8506e8be71c91d0656558f8529bd3b" Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.932608 4775 generic.go:334] "Generic (PLEG): container finished" podID="7da7ce8b-402a-40f7-a9ce-340524ad8573" containerID="b57b29e11c242340693c81affc2d61431477f8ba75c091165a111504297edd4f" exitCode=0 Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.932630 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" event={"ID":"7da7ce8b-402a-40f7-a9ce-340524ad8573","Type":"ContainerDied","Data":"b57b29e11c242340693c81affc2d61431477f8ba75c091165a111504297edd4f"} Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.932647 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.932651 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f88f547f-kjxrc" event={"ID":"7da7ce8b-402a-40f7-a9ce-340524ad8573","Type":"ContainerDied","Data":"2e2beab81a4014c294e6052e86efc7d9c966045a25b61fae9abde1855c0e91eb"} Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.945315 4775 scope.go:117] "RemoveContainer" containerID="ae81d0a5c5add9cc57df7dd09c08b1e07e8506e8be71c91d0656558f8529bd3b" Mar 21 04:53:34 crc kubenswrapper[4775]: E0321 04:53:34.945748 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae81d0a5c5add9cc57df7dd09c08b1e07e8506e8be71c91d0656558f8529bd3b\": container with ID starting with ae81d0a5c5add9cc57df7dd09c08b1e07e8506e8be71c91d0656558f8529bd3b not found: ID does not exist" containerID="ae81d0a5c5add9cc57df7dd09c08b1e07e8506e8be71c91d0656558f8529bd3b" Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.945792 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae81d0a5c5add9cc57df7dd09c08b1e07e8506e8be71c91d0656558f8529bd3b"} err="failed to get container status \"ae81d0a5c5add9cc57df7dd09c08b1e07e8506e8be71c91d0656558f8529bd3b\": rpc error: code = NotFound desc = could not find container \"ae81d0a5c5add9cc57df7dd09c08b1e07e8506e8be71c91d0656558f8529bd3b\": container with ID starting with ae81d0a5c5add9cc57df7dd09c08b1e07e8506e8be71c91d0656558f8529bd3b not found: ID does not exist" Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.945817 4775 scope.go:117] "RemoveContainer" containerID="b57b29e11c242340693c81affc2d61431477f8ba75c091165a111504297edd4f" Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.965411 4775 scope.go:117] "RemoveContainer" containerID="b57b29e11c242340693c81affc2d61431477f8ba75c091165a111504297edd4f" Mar 21 04:53:34 crc kubenswrapper[4775]: E0321 04:53:34.965810 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b57b29e11c242340693c81affc2d61431477f8ba75c091165a111504297edd4f\": container with ID starting with b57b29e11c242340693c81affc2d61431477f8ba75c091165a111504297edd4f not found: ID does not exist" containerID="b57b29e11c242340693c81affc2d61431477f8ba75c091165a111504297edd4f" Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.965846 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b57b29e11c242340693c81affc2d61431477f8ba75c091165a111504297edd4f"} err="failed to get container status \"b57b29e11c242340693c81affc2d61431477f8ba75c091165a111504297edd4f\": rpc error: code = NotFound desc = could not find container \"b57b29e11c242340693c81affc2d61431477f8ba75c091165a111504297edd4f\": container with ID starting with b57b29e11c242340693c81affc2d61431477f8ba75c091165a111504297edd4f not found: ID does not exist" Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.967655 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6f88f547f-kjxrc"] Mar 21 04:53:34 crc kubenswrapper[4775]: I0321 04:53:34.974387 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6f88f547f-kjxrc"] Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.005005 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65v5z\" (UniqueName: \"kubernetes.io/projected/1882b1c5-2465-40df-9306-e5792e0d9f2f-kube-api-access-65v5z\") pod \"1882b1c5-2465-40df-9306-e5792e0d9f2f\" (UID: \"1882b1c5-2465-40df-9306-e5792e0d9f2f\") " Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.005050 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1882b1c5-2465-40df-9306-e5792e0d9f2f-client-ca\") pod \"1882b1c5-2465-40df-9306-e5792e0d9f2f\" (UID: \"1882b1c5-2465-40df-9306-e5792e0d9f2f\") " Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.005090 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1882b1c5-2465-40df-9306-e5792e0d9f2f-config\") pod \"1882b1c5-2465-40df-9306-e5792e0d9f2f\" (UID: \"1882b1c5-2465-40df-9306-e5792e0d9f2f\") " Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.005210 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1882b1c5-2465-40df-9306-e5792e0d9f2f-serving-cert\") pod \"1882b1c5-2465-40df-9306-e5792e0d9f2f\" (UID: \"1882b1c5-2465-40df-9306-e5792e0d9f2f\") " Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.005480 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7da7ce8b-402a-40f7-a9ce-340524ad8573-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.005498 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7da7ce8b-402a-40f7-a9ce-340524ad8573-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.005509 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7da7ce8b-402a-40f7-a9ce-340524ad8573-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.005523 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flbhf\" (UniqueName: \"kubernetes.io/projected/7da7ce8b-402a-40f7-a9ce-340524ad8573-kube-api-access-flbhf\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.005536 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da7ce8b-402a-40f7-a9ce-340524ad8573-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.005795 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1882b1c5-2465-40df-9306-e5792e0d9f2f-client-ca" (OuterVolumeSpecName: "client-ca") pod "1882b1c5-2465-40df-9306-e5792e0d9f2f" (UID: "1882b1c5-2465-40df-9306-e5792e0d9f2f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.006177 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1882b1c5-2465-40df-9306-e5792e0d9f2f-config" (OuterVolumeSpecName: "config") pod "1882b1c5-2465-40df-9306-e5792e0d9f2f" (UID: "1882b1c5-2465-40df-9306-e5792e0d9f2f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.007837 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1882b1c5-2465-40df-9306-e5792e0d9f2f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1882b1c5-2465-40df-9306-e5792e0d9f2f" (UID: "1882b1c5-2465-40df-9306-e5792e0d9f2f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.007961 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1882b1c5-2465-40df-9306-e5792e0d9f2f-kube-api-access-65v5z" (OuterVolumeSpecName: "kube-api-access-65v5z") pod "1882b1c5-2465-40df-9306-e5792e0d9f2f" (UID: "1882b1c5-2465-40df-9306-e5792e0d9f2f"). InnerVolumeSpecName "kube-api-access-65v5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.106407 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65v5z\" (UniqueName: \"kubernetes.io/projected/1882b1c5-2465-40df-9306-e5792e0d9f2f-kube-api-access-65v5z\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.106493 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1882b1c5-2465-40df-9306-e5792e0d9f2f-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.106516 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1882b1c5-2465-40df-9306-e5792e0d9f2f-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.106536 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1882b1c5-2465-40df-9306-e5792e0d9f2f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.263082 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z"] Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.266087 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d6c87d469-rgj5z"] Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.668039 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1882b1c5-2465-40df-9306-e5792e0d9f2f" path="/var/lib/kubelet/pods/1882b1c5-2465-40df-9306-e5792e0d9f2f/volumes" Mar 21 04:53:35 crc kubenswrapper[4775]: I0321 04:53:35.668746 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7da7ce8b-402a-40f7-a9ce-340524ad8573" path="/var/lib/kubelet/pods/7da7ce8b-402a-40f7-a9ce-340524ad8573/volumes" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.021270 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5b5b8879f4-snbl7"] Mar 21 04:53:36 crc kubenswrapper[4775]: E0321 04:53:36.021534 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7da7ce8b-402a-40f7-a9ce-340524ad8573" containerName="controller-manager" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.021549 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7da7ce8b-402a-40f7-a9ce-340524ad8573" containerName="controller-manager" Mar 21 04:53:36 crc kubenswrapper[4775]: E0321 04:53:36.021561 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.021567 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 21 04:53:36 crc kubenswrapper[4775]: E0321 04:53:36.021580 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1882b1c5-2465-40df-9306-e5792e0d9f2f" containerName="route-controller-manager" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.021588 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1882b1c5-2465-40df-9306-e5792e0d9f2f" containerName="route-controller-manager" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.021698 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7da7ce8b-402a-40f7-a9ce-340524ad8573" containerName="controller-manager" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.021717 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.021724 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1882b1c5-2465-40df-9306-e5792e0d9f2f" containerName="route-controller-manager" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.022049 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.025438 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.025550 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.025644 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.025654 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.026862 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.027193 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.029731 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq"] Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.030454 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.032839 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.033222 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.033400 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.033454 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.033620 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.033704 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.034369 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq"] Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.036908 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.044676 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5b5b8879f4-snbl7"] Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.117731 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/28d4e02d-a2fa-4465-936b-df5f7ef22086-proxy-ca-bundles\") pod \"controller-manager-5b5b8879f4-snbl7\" (UID: \"28d4e02d-a2fa-4465-936b-df5f7ef22086\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.117778 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28d4e02d-a2fa-4465-936b-df5f7ef22086-config\") pod \"controller-manager-5b5b8879f4-snbl7\" (UID: \"28d4e02d-a2fa-4465-936b-df5f7ef22086\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.117803 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxv79\" (UniqueName: \"kubernetes.io/projected/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-kube-api-access-mxv79\") pod \"route-controller-manager-555c6b4d6b-jh8tq\" (UID: \"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.117825 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/28d4e02d-a2fa-4465-936b-df5f7ef22086-client-ca\") pod \"controller-manager-5b5b8879f4-snbl7\" (UID: \"28d4e02d-a2fa-4465-936b-df5f7ef22086\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.117945 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-serving-cert\") pod \"route-controller-manager-555c6b4d6b-jh8tq\" (UID: \"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.118021 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-config\") pod \"route-controller-manager-555c6b4d6b-jh8tq\" (UID: \"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.118055 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28d4e02d-a2fa-4465-936b-df5f7ef22086-serving-cert\") pod \"controller-manager-5b5b8879f4-snbl7\" (UID: \"28d4e02d-a2fa-4465-936b-df5f7ef22086\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.118079 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67jrx\" (UniqueName: \"kubernetes.io/projected/28d4e02d-a2fa-4465-936b-df5f7ef22086-kube-api-access-67jrx\") pod \"controller-manager-5b5b8879f4-snbl7\" (UID: \"28d4e02d-a2fa-4465-936b-df5f7ef22086\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.118102 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-client-ca\") pod \"route-controller-manager-555c6b4d6b-jh8tq\" (UID: \"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.219790 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/28d4e02d-a2fa-4465-936b-df5f7ef22086-proxy-ca-bundles\") pod \"controller-manager-5b5b8879f4-snbl7\" (UID: \"28d4e02d-a2fa-4465-936b-df5f7ef22086\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.219841 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28d4e02d-a2fa-4465-936b-df5f7ef22086-config\") pod \"controller-manager-5b5b8879f4-snbl7\" (UID: \"28d4e02d-a2fa-4465-936b-df5f7ef22086\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.219863 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxv79\" (UniqueName: \"kubernetes.io/projected/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-kube-api-access-mxv79\") pod \"route-controller-manager-555c6b4d6b-jh8tq\" (UID: \"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.219885 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/28d4e02d-a2fa-4465-936b-df5f7ef22086-client-ca\") pod \"controller-manager-5b5b8879f4-snbl7\" (UID: \"28d4e02d-a2fa-4465-936b-df5f7ef22086\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.219910 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-serving-cert\") pod \"route-controller-manager-555c6b4d6b-jh8tq\" (UID: \"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.219943 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-config\") pod \"route-controller-manager-555c6b4d6b-jh8tq\" (UID: \"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.219984 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28d4e02d-a2fa-4465-936b-df5f7ef22086-serving-cert\") pod \"controller-manager-5b5b8879f4-snbl7\" (UID: \"28d4e02d-a2fa-4465-936b-df5f7ef22086\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.220019 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67jrx\" (UniqueName: \"kubernetes.io/projected/28d4e02d-a2fa-4465-936b-df5f7ef22086-kube-api-access-67jrx\") pod \"controller-manager-5b5b8879f4-snbl7\" (UID: \"28d4e02d-a2fa-4465-936b-df5f7ef22086\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.220045 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-client-ca\") pod \"route-controller-manager-555c6b4d6b-jh8tq\" (UID: \"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.221217 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-client-ca\") pod \"route-controller-manager-555c6b4d6b-jh8tq\" (UID: \"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.221283 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/28d4e02d-a2fa-4465-936b-df5f7ef22086-client-ca\") pod \"controller-manager-5b5b8879f4-snbl7\" (UID: \"28d4e02d-a2fa-4465-936b-df5f7ef22086\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.221403 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/28d4e02d-a2fa-4465-936b-df5f7ef22086-proxy-ca-bundles\") pod \"controller-manager-5b5b8879f4-snbl7\" (UID: \"28d4e02d-a2fa-4465-936b-df5f7ef22086\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.221505 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28d4e02d-a2fa-4465-936b-df5f7ef22086-config\") pod \"controller-manager-5b5b8879f4-snbl7\" (UID: \"28d4e02d-a2fa-4465-936b-df5f7ef22086\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.222501 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-config\") pod \"route-controller-manager-555c6b4d6b-jh8tq\" (UID: \"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.224050 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-serving-cert\") pod \"route-controller-manager-555c6b4d6b-jh8tq\" (UID: \"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.232665 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28d4e02d-a2fa-4465-936b-df5f7ef22086-serving-cert\") pod \"controller-manager-5b5b8879f4-snbl7\" (UID: \"28d4e02d-a2fa-4465-936b-df5f7ef22086\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.236331 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxv79\" (UniqueName: \"kubernetes.io/projected/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-kube-api-access-mxv79\") pod \"route-controller-manager-555c6b4d6b-jh8tq\" (UID: \"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.238865 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67jrx\" (UniqueName: \"kubernetes.io/projected/28d4e02d-a2fa-4465-936b-df5f7ef22086-kube-api-access-67jrx\") pod \"controller-manager-5b5b8879f4-snbl7\" (UID: \"28d4e02d-a2fa-4465-936b-df5f7ef22086\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.347660 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.363158 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.768144 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5b5b8879f4-snbl7"] Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.796990 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq"] Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.952924 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" event={"ID":"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb","Type":"ContainerStarted","Data":"56c232d6ac9969429c18b124a338e1b6499df231f55af06b4690d442d2b765b2"} Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.954907 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" event={"ID":"28d4e02d-a2fa-4465-936b-df5f7ef22086","Type":"ContainerStarted","Data":"87509b0780118b1a8a1bce50fbeee0c0687a23effaec2af6b41e83b608d4b8d9"} Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.955164 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.957092 4775 patch_prober.go:28] interesting pod/controller-manager-5b5b8879f4-snbl7 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" start-of-body= Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.957159 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" podUID="28d4e02d-a2fa-4465-936b-df5f7ef22086" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" Mar 21 04:53:36 crc kubenswrapper[4775]: I0321 04:53:36.975234 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" podStartSLOduration=2.9752173859999997 podStartE2EDuration="2.975217386s" podCreationTimestamp="2026-03-21 04:53:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:53:36.970920411 +0000 UTC m=+369.947384035" watchObservedRunningTime="2026-03-21 04:53:36.975217386 +0000 UTC m=+369.951681010" Mar 21 04:53:37 crc kubenswrapper[4775]: I0321 04:53:37.961626 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" event={"ID":"28d4e02d-a2fa-4465-936b-df5f7ef22086","Type":"ContainerStarted","Data":"20c1b7542354b374133f523fc8f60ffcbd5359eb5d860eb5669e6229862d91e5"} Mar 21 04:53:37 crc kubenswrapper[4775]: I0321 04:53:37.963671 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" event={"ID":"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb","Type":"ContainerStarted","Data":"2115ff11bda799954a12857b7ea38cd8c929ef48a09eaaeb6e02765150d49cb8"} Mar 21 04:53:37 crc kubenswrapper[4775]: I0321 04:53:37.964022 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" Mar 21 04:53:37 crc kubenswrapper[4775]: I0321 04:53:37.966371 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:37 crc kubenswrapper[4775]: I0321 04:53:37.968178 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" Mar 21 04:53:37 crc kubenswrapper[4775]: I0321 04:53:37.979831 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" podStartSLOduration=3.979815002 podStartE2EDuration="3.979815002s" podCreationTimestamp="2026-03-21 04:53:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:53:37.97835616 +0000 UTC m=+370.954819784" watchObservedRunningTime="2026-03-21 04:53:37.979815002 +0000 UTC m=+370.956278626" Mar 21 04:53:43 crc kubenswrapper[4775]: I0321 04:53:43.997663 4775 generic.go:334] "Generic (PLEG): container finished" podID="dfbaac71-f99c-4373-a469-f2e5dd0ee632" containerID="e6613d34aa689f2c33b46ca7fb2fff89fd26737df8f8eb26a52c2030909ae41f" exitCode=0 Mar 21 04:53:43 crc kubenswrapper[4775]: I0321 04:53:43.997728 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" event={"ID":"dfbaac71-f99c-4373-a469-f2e5dd0ee632","Type":"ContainerDied","Data":"e6613d34aa689f2c33b46ca7fb2fff89fd26737df8f8eb26a52c2030909ae41f"} Mar 21 04:53:43 crc kubenswrapper[4775]: I0321 04:53:43.999221 4775 scope.go:117] "RemoveContainer" containerID="e6613d34aa689f2c33b46ca7fb2fff89fd26737df8f8eb26a52c2030909ae41f" Mar 21 04:53:45 crc kubenswrapper[4775]: I0321 04:53:45.013139 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" event={"ID":"dfbaac71-f99c-4373-a469-f2e5dd0ee632","Type":"ContainerStarted","Data":"53e93eb8775f797ec5701d0054608e0cc06d124aa1940ad067f4c33b12a82fa2"} Mar 21 04:53:45 crc kubenswrapper[4775]: I0321 04:53:45.013779 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" Mar 21 04:53:45 crc kubenswrapper[4775]: I0321 04:53:45.017792 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" Mar 21 04:53:54 crc kubenswrapper[4775]: I0321 04:53:54.949272 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5b5b8879f4-snbl7"] Mar 21 04:53:54 crc kubenswrapper[4775]: I0321 04:53:54.949973 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" podUID="28d4e02d-a2fa-4465-936b-df5f7ef22086" containerName="controller-manager" containerID="cri-o://20c1b7542354b374133f523fc8f60ffcbd5359eb5d860eb5669e6229862d91e5" gracePeriod=30 Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.029870 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq"] Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.030276 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" podUID="e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb" containerName="route-controller-manager" containerID="cri-o://2115ff11bda799954a12857b7ea38cd8c929ef48a09eaaeb6e02765150d49cb8" gracePeriod=30 Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.070399 4775 generic.go:334] "Generic (PLEG): container finished" podID="28d4e02d-a2fa-4465-936b-df5f7ef22086" containerID="20c1b7542354b374133f523fc8f60ffcbd5359eb5d860eb5669e6229862d91e5" exitCode=0 Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.070437 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" event={"ID":"28d4e02d-a2fa-4465-936b-df5f7ef22086","Type":"ContainerDied","Data":"20c1b7542354b374133f523fc8f60ffcbd5359eb5d860eb5669e6229862d91e5"} Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.478187 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.537186 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.555456 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-client-ca\") pod \"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb\" (UID: \"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb\") " Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.555501 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-serving-cert\") pod \"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb\" (UID: \"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb\") " Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.555538 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxv79\" (UniqueName: \"kubernetes.io/projected/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-kube-api-access-mxv79\") pod \"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb\" (UID: \"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb\") " Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.555570 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-config\") pod \"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb\" (UID: \"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb\") " Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.556333 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-client-ca" (OuterVolumeSpecName: "client-ca") pod "e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb" (UID: "e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.556345 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-config" (OuterVolumeSpecName: "config") pod "e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb" (UID: "e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.556558 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.556576 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.563006 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-kube-api-access-mxv79" (OuterVolumeSpecName: "kube-api-access-mxv79") pod "e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb" (UID: "e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb"). InnerVolumeSpecName "kube-api-access-mxv79". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.563349 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb" (UID: "e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.657073 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28d4e02d-a2fa-4465-936b-df5f7ef22086-config\") pod \"28d4e02d-a2fa-4465-936b-df5f7ef22086\" (UID: \"28d4e02d-a2fa-4465-936b-df5f7ef22086\") " Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.657133 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/28d4e02d-a2fa-4465-936b-df5f7ef22086-proxy-ca-bundles\") pod \"28d4e02d-a2fa-4465-936b-df5f7ef22086\" (UID: \"28d4e02d-a2fa-4465-936b-df5f7ef22086\") " Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.657159 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/28d4e02d-a2fa-4465-936b-df5f7ef22086-client-ca\") pod \"28d4e02d-a2fa-4465-936b-df5f7ef22086\" (UID: \"28d4e02d-a2fa-4465-936b-df5f7ef22086\") " Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.657226 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28d4e02d-a2fa-4465-936b-df5f7ef22086-serving-cert\") pod \"28d4e02d-a2fa-4465-936b-df5f7ef22086\" (UID: \"28d4e02d-a2fa-4465-936b-df5f7ef22086\") " Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.657263 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67jrx\" (UniqueName: \"kubernetes.io/projected/28d4e02d-a2fa-4465-936b-df5f7ef22086-kube-api-access-67jrx\") pod \"28d4e02d-a2fa-4465-936b-df5f7ef22086\" (UID: \"28d4e02d-a2fa-4465-936b-df5f7ef22086\") " Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.657540 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.657557 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxv79\" (UniqueName: \"kubernetes.io/projected/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb-kube-api-access-mxv79\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.657875 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28d4e02d-a2fa-4465-936b-df5f7ef22086-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "28d4e02d-a2fa-4465-936b-df5f7ef22086" (UID: "28d4e02d-a2fa-4465-936b-df5f7ef22086"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.657958 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28d4e02d-a2fa-4465-936b-df5f7ef22086-config" (OuterVolumeSpecName: "config") pod "28d4e02d-a2fa-4465-936b-df5f7ef22086" (UID: "28d4e02d-a2fa-4465-936b-df5f7ef22086"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.658513 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28d4e02d-a2fa-4465-936b-df5f7ef22086-client-ca" (OuterVolumeSpecName: "client-ca") pod "28d4e02d-a2fa-4465-936b-df5f7ef22086" (UID: "28d4e02d-a2fa-4465-936b-df5f7ef22086"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.660065 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28d4e02d-a2fa-4465-936b-df5f7ef22086-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "28d4e02d-a2fa-4465-936b-df5f7ef22086" (UID: "28d4e02d-a2fa-4465-936b-df5f7ef22086"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.660262 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28d4e02d-a2fa-4465-936b-df5f7ef22086-kube-api-access-67jrx" (OuterVolumeSpecName: "kube-api-access-67jrx") pod "28d4e02d-a2fa-4465-936b-df5f7ef22086" (UID: "28d4e02d-a2fa-4465-936b-df5f7ef22086"). InnerVolumeSpecName "kube-api-access-67jrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.759476 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28d4e02d-a2fa-4465-936b-df5f7ef22086-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.759529 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67jrx\" (UniqueName: \"kubernetes.io/projected/28d4e02d-a2fa-4465-936b-df5f7ef22086-kube-api-access-67jrx\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.759550 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28d4e02d-a2fa-4465-936b-df5f7ef22086-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.759567 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/28d4e02d-a2fa-4465-936b-df5f7ef22086-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:55 crc kubenswrapper[4775]: I0321 04:53:55.759586 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/28d4e02d-a2fa-4465-936b-df5f7ef22086-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.033519 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-77fc8489b9-p4xpt"] Mar 21 04:53:56 crc kubenswrapper[4775]: E0321 04:53:56.033705 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb" containerName="route-controller-manager" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.033717 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb" containerName="route-controller-manager" Mar 21 04:53:56 crc kubenswrapper[4775]: E0321 04:53:56.033732 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d4e02d-a2fa-4465-936b-df5f7ef22086" containerName="controller-manager" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.033740 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d4e02d-a2fa-4465-936b-df5f7ef22086" containerName="controller-manager" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.033834 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb" containerName="route-controller-manager" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.033846 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="28d4e02d-a2fa-4465-936b-df5f7ef22086" containerName="controller-manager" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.034202 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.043314 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-77fc8489b9-p4xpt"] Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.079975 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.079961 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b5b8879f4-snbl7" event={"ID":"28d4e02d-a2fa-4465-936b-df5f7ef22086","Type":"ContainerDied","Data":"87509b0780118b1a8a1bce50fbeee0c0687a23effaec2af6b41e83b608d4b8d9"} Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.080196 4775 scope.go:117] "RemoveContainer" containerID="20c1b7542354b374133f523fc8f60ffcbd5359eb5d860eb5669e6229862d91e5" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.081591 4775 generic.go:334] "Generic (PLEG): container finished" podID="e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb" containerID="2115ff11bda799954a12857b7ea38cd8c929ef48a09eaaeb6e02765150d49cb8" exitCode=0 Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.081640 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.081634 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" event={"ID":"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb","Type":"ContainerDied","Data":"2115ff11bda799954a12857b7ea38cd8c929ef48a09eaaeb6e02765150d49cb8"} Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.081766 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq" event={"ID":"e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb","Type":"ContainerDied","Data":"56c232d6ac9969429c18b124a338e1b6499df231f55af06b4690d442d2b765b2"} Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.101186 4775 scope.go:117] "RemoveContainer" containerID="2115ff11bda799954a12857b7ea38cd8c929ef48a09eaaeb6e02765150d49cb8" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.116023 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5b5b8879f4-snbl7"] Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.117857 4775 scope.go:117] "RemoveContainer" containerID="2115ff11bda799954a12857b7ea38cd8c929ef48a09eaaeb6e02765150d49cb8" Mar 21 04:53:56 crc kubenswrapper[4775]: E0321 04:53:56.118388 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2115ff11bda799954a12857b7ea38cd8c929ef48a09eaaeb6e02765150d49cb8\": container with ID starting with 2115ff11bda799954a12857b7ea38cd8c929ef48a09eaaeb6e02765150d49cb8 not found: ID does not exist" containerID="2115ff11bda799954a12857b7ea38cd8c929ef48a09eaaeb6e02765150d49cb8" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.118440 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2115ff11bda799954a12857b7ea38cd8c929ef48a09eaaeb6e02765150d49cb8"} err="failed to get container status \"2115ff11bda799954a12857b7ea38cd8c929ef48a09eaaeb6e02765150d49cb8\": rpc error: code = NotFound desc = could not find container \"2115ff11bda799954a12857b7ea38cd8c929ef48a09eaaeb6e02765150d49cb8\": container with ID starting with 2115ff11bda799954a12857b7ea38cd8c929ef48a09eaaeb6e02765150d49cb8 not found: ID does not exist" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.120344 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5b5b8879f4-snbl7"] Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.130225 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq"] Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.133303 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-555c6b4d6b-jh8tq"] Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.164364 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41534621-ccd0-4e69-a13c-e6828feebe17-serving-cert\") pod \"controller-manager-77fc8489b9-p4xpt\" (UID: \"41534621-ccd0-4e69-a13c-e6828feebe17\") " pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.164422 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41534621-ccd0-4e69-a13c-e6828feebe17-config\") pod \"controller-manager-77fc8489b9-p4xpt\" (UID: \"41534621-ccd0-4e69-a13c-e6828feebe17\") " pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.164448 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41534621-ccd0-4e69-a13c-e6828feebe17-client-ca\") pod \"controller-manager-77fc8489b9-p4xpt\" (UID: \"41534621-ccd0-4e69-a13c-e6828feebe17\") " pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.164486 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s48s5\" (UniqueName: \"kubernetes.io/projected/41534621-ccd0-4e69-a13c-e6828feebe17-kube-api-access-s48s5\") pod \"controller-manager-77fc8489b9-p4xpt\" (UID: \"41534621-ccd0-4e69-a13c-e6828feebe17\") " pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.164520 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/41534621-ccd0-4e69-a13c-e6828feebe17-proxy-ca-bundles\") pod \"controller-manager-77fc8489b9-p4xpt\" (UID: \"41534621-ccd0-4e69-a13c-e6828feebe17\") " pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.265520 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41534621-ccd0-4e69-a13c-e6828feebe17-serving-cert\") pod \"controller-manager-77fc8489b9-p4xpt\" (UID: \"41534621-ccd0-4e69-a13c-e6828feebe17\") " pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.265695 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41534621-ccd0-4e69-a13c-e6828feebe17-config\") pod \"controller-manager-77fc8489b9-p4xpt\" (UID: \"41534621-ccd0-4e69-a13c-e6828feebe17\") " pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.265752 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41534621-ccd0-4e69-a13c-e6828feebe17-client-ca\") pod \"controller-manager-77fc8489b9-p4xpt\" (UID: \"41534621-ccd0-4e69-a13c-e6828feebe17\") " pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.265796 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s48s5\" (UniqueName: \"kubernetes.io/projected/41534621-ccd0-4e69-a13c-e6828feebe17-kube-api-access-s48s5\") pod \"controller-manager-77fc8489b9-p4xpt\" (UID: \"41534621-ccd0-4e69-a13c-e6828feebe17\") " pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.265860 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/41534621-ccd0-4e69-a13c-e6828feebe17-proxy-ca-bundles\") pod \"controller-manager-77fc8489b9-p4xpt\" (UID: \"41534621-ccd0-4e69-a13c-e6828feebe17\") " pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.266825 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41534621-ccd0-4e69-a13c-e6828feebe17-client-ca\") pod \"controller-manager-77fc8489b9-p4xpt\" (UID: \"41534621-ccd0-4e69-a13c-e6828feebe17\") " pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.267451 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41534621-ccd0-4e69-a13c-e6828feebe17-config\") pod \"controller-manager-77fc8489b9-p4xpt\" (UID: \"41534621-ccd0-4e69-a13c-e6828feebe17\") " pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.267540 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/41534621-ccd0-4e69-a13c-e6828feebe17-proxy-ca-bundles\") pod \"controller-manager-77fc8489b9-p4xpt\" (UID: \"41534621-ccd0-4e69-a13c-e6828feebe17\") " pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.271130 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41534621-ccd0-4e69-a13c-e6828feebe17-serving-cert\") pod \"controller-manager-77fc8489b9-p4xpt\" (UID: \"41534621-ccd0-4e69-a13c-e6828feebe17\") " pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.282681 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s48s5\" (UniqueName: \"kubernetes.io/projected/41534621-ccd0-4e69-a13c-e6828feebe17-kube-api-access-s48s5\") pod \"controller-manager-77fc8489b9-p4xpt\" (UID: \"41534621-ccd0-4e69-a13c-e6828feebe17\") " pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.347434 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:53:56 crc kubenswrapper[4775]: I0321 04:53:56.746617 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-77fc8489b9-p4xpt"] Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.034746 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv"] Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.035481 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.036824 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.037186 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.037551 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.037856 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.038153 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.041421 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.045512 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv"] Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.088695 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" event={"ID":"41534621-ccd0-4e69-a13c-e6828feebe17","Type":"ContainerStarted","Data":"5104548b211c2e16bc96a01a5e3e4f9c0385383b17796ee059444b9f78c92a94"} Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.088763 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" event={"ID":"41534621-ccd0-4e69-a13c-e6828feebe17","Type":"ContainerStarted","Data":"69122f1ec9bc520b92bd94027f2fae1d2dd0387607089568975168c52c43b3a7"} Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.089202 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.103788 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.112629 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" podStartSLOduration=3.112613185 podStartE2EDuration="3.112613185s" podCreationTimestamp="2026-03-21 04:53:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:53:57.109787113 +0000 UTC m=+390.086250757" watchObservedRunningTime="2026-03-21 04:53:57.112613185 +0000 UTC m=+390.089076809" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.180911 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4dwd\" (UniqueName: \"kubernetes.io/projected/951da48c-671e-4004-85f0-2137949a3a2d-kube-api-access-d4dwd\") pod \"route-controller-manager-7d78b4c7b9-244bv\" (UID: \"951da48c-671e-4004-85f0-2137949a3a2d\") " pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.181279 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/951da48c-671e-4004-85f0-2137949a3a2d-config\") pod \"route-controller-manager-7d78b4c7b9-244bv\" (UID: \"951da48c-671e-4004-85f0-2137949a3a2d\") " pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.181368 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/951da48c-671e-4004-85f0-2137949a3a2d-client-ca\") pod \"route-controller-manager-7d78b4c7b9-244bv\" (UID: \"951da48c-671e-4004-85f0-2137949a3a2d\") " pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.181388 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/951da48c-671e-4004-85f0-2137949a3a2d-serving-cert\") pod \"route-controller-manager-7d78b4c7b9-244bv\" (UID: \"951da48c-671e-4004-85f0-2137949a3a2d\") " pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.282190 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/951da48c-671e-4004-85f0-2137949a3a2d-client-ca\") pod \"route-controller-manager-7d78b4c7b9-244bv\" (UID: \"951da48c-671e-4004-85f0-2137949a3a2d\") " pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.282247 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/951da48c-671e-4004-85f0-2137949a3a2d-serving-cert\") pod \"route-controller-manager-7d78b4c7b9-244bv\" (UID: \"951da48c-671e-4004-85f0-2137949a3a2d\") " pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.282293 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4dwd\" (UniqueName: \"kubernetes.io/projected/951da48c-671e-4004-85f0-2137949a3a2d-kube-api-access-d4dwd\") pod \"route-controller-manager-7d78b4c7b9-244bv\" (UID: \"951da48c-671e-4004-85f0-2137949a3a2d\") " pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.282332 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/951da48c-671e-4004-85f0-2137949a3a2d-config\") pod \"route-controller-manager-7d78b4c7b9-244bv\" (UID: \"951da48c-671e-4004-85f0-2137949a3a2d\") " pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.283447 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/951da48c-671e-4004-85f0-2137949a3a2d-client-ca\") pod \"route-controller-manager-7d78b4c7b9-244bv\" (UID: \"951da48c-671e-4004-85f0-2137949a3a2d\") " pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.283541 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/951da48c-671e-4004-85f0-2137949a3a2d-config\") pod \"route-controller-manager-7d78b4c7b9-244bv\" (UID: \"951da48c-671e-4004-85f0-2137949a3a2d\") " pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.287352 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/951da48c-671e-4004-85f0-2137949a3a2d-serving-cert\") pod \"route-controller-manager-7d78b4c7b9-244bv\" (UID: \"951da48c-671e-4004-85f0-2137949a3a2d\") " pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.298316 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4dwd\" (UniqueName: \"kubernetes.io/projected/951da48c-671e-4004-85f0-2137949a3a2d-kube-api-access-d4dwd\") pod \"route-controller-manager-7d78b4c7b9-244bv\" (UID: \"951da48c-671e-4004-85f0-2137949a3a2d\") " pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.356687 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.669374 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28d4e02d-a2fa-4465-936b-df5f7ef22086" path="/var/lib/kubelet/pods/28d4e02d-a2fa-4465-936b-df5f7ef22086/volumes" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.670154 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb" path="/var/lib/kubelet/pods/e3e7cd58-51b5-4a7f-b374-fe94b9b0c1eb/volumes" Mar 21 04:53:57 crc kubenswrapper[4775]: I0321 04:53:57.793948 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv"] Mar 21 04:53:57 crc kubenswrapper[4775]: W0321 04:53:57.800337 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod951da48c_671e_4004_85f0_2137949a3a2d.slice/crio-aa646cd6ba5a44de57d221735dcb1de71958eda8b620088cc7ef6124aed0413e WatchSource:0}: Error finding container aa646cd6ba5a44de57d221735dcb1de71958eda8b620088cc7ef6124aed0413e: Status 404 returned error can't find the container with id aa646cd6ba5a44de57d221735dcb1de71958eda8b620088cc7ef6124aed0413e Mar 21 04:53:58 crc kubenswrapper[4775]: I0321 04:53:58.099316 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" event={"ID":"951da48c-671e-4004-85f0-2137949a3a2d","Type":"ContainerStarted","Data":"3980fc5ac2c84abc585c64be5fe0e20cdc2ee02408d0873f172d592ea567b97a"} Mar 21 04:53:58 crc kubenswrapper[4775]: I0321 04:53:58.099383 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" event={"ID":"951da48c-671e-4004-85f0-2137949a3a2d","Type":"ContainerStarted","Data":"aa646cd6ba5a44de57d221735dcb1de71958eda8b620088cc7ef6124aed0413e"} Mar 21 04:53:58 crc kubenswrapper[4775]: I0321 04:53:58.099547 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" Mar 21 04:53:58 crc kubenswrapper[4775]: I0321 04:53:58.335779 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" Mar 21 04:53:58 crc kubenswrapper[4775]: I0321 04:53:58.355958 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" podStartSLOduration=3.355935454 podStartE2EDuration="3.355935454s" podCreationTimestamp="2026-03-21 04:53:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:53:58.116677296 +0000 UTC m=+391.093140920" watchObservedRunningTime="2026-03-21 04:53:58.355935454 +0000 UTC m=+391.332399078" Mar 21 04:54:00 crc kubenswrapper[4775]: I0321 04:54:00.158587 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567814-fvs8k"] Mar 21 04:54:00 crc kubenswrapper[4775]: I0321 04:54:00.160001 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567814-fvs8k" Mar 21 04:54:00 crc kubenswrapper[4775]: I0321 04:54:00.222325 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 04:54:00 crc kubenswrapper[4775]: I0321 04:54:00.223956 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 04:54:00 crc kubenswrapper[4775]: I0321 04:54:00.224505 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 04:54:00 crc kubenswrapper[4775]: I0321 04:54:00.234168 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567814-fvs8k"] Mar 21 04:54:00 crc kubenswrapper[4775]: I0321 04:54:00.323080 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfgpl\" (UniqueName: \"kubernetes.io/projected/e8f9e3cb-7583-4349-98dc-422a7880fe5e-kube-api-access-gfgpl\") pod \"auto-csr-approver-29567814-fvs8k\" (UID: \"e8f9e3cb-7583-4349-98dc-422a7880fe5e\") " pod="openshift-infra/auto-csr-approver-29567814-fvs8k" Mar 21 04:54:00 crc kubenswrapper[4775]: I0321 04:54:00.423861 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfgpl\" (UniqueName: \"kubernetes.io/projected/e8f9e3cb-7583-4349-98dc-422a7880fe5e-kube-api-access-gfgpl\") pod \"auto-csr-approver-29567814-fvs8k\" (UID: \"e8f9e3cb-7583-4349-98dc-422a7880fe5e\") " pod="openshift-infra/auto-csr-approver-29567814-fvs8k" Mar 21 04:54:00 crc kubenswrapper[4775]: I0321 04:54:00.443669 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfgpl\" (UniqueName: \"kubernetes.io/projected/e8f9e3cb-7583-4349-98dc-422a7880fe5e-kube-api-access-gfgpl\") pod \"auto-csr-approver-29567814-fvs8k\" (UID: \"e8f9e3cb-7583-4349-98dc-422a7880fe5e\") " pod="openshift-infra/auto-csr-approver-29567814-fvs8k" Mar 21 04:54:00 crc kubenswrapper[4775]: I0321 04:54:00.573000 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567814-fvs8k" Mar 21 04:54:01 crc kubenswrapper[4775]: I0321 04:54:01.004888 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567814-fvs8k"] Mar 21 04:54:01 crc kubenswrapper[4775]: I0321 04:54:01.118787 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567814-fvs8k" event={"ID":"e8f9e3cb-7583-4349-98dc-422a7880fe5e","Type":"ContainerStarted","Data":"bef13ceeba8d06c136483de90c1fc8ff328392147f003ed93dfc860c9a4863ee"} Mar 21 04:54:03 crc kubenswrapper[4775]: I0321 04:54:03.131852 4775 generic.go:334] "Generic (PLEG): container finished" podID="e8f9e3cb-7583-4349-98dc-422a7880fe5e" containerID="55e31feec3221e168e396a1b14aac3afcd926bdfa911a175f057a586cb60827a" exitCode=0 Mar 21 04:54:03 crc kubenswrapper[4775]: I0321 04:54:03.131903 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567814-fvs8k" event={"ID":"e8f9e3cb-7583-4349-98dc-422a7880fe5e","Type":"ContainerDied","Data":"55e31feec3221e168e396a1b14aac3afcd926bdfa911a175f057a586cb60827a"} Mar 21 04:54:04 crc kubenswrapper[4775]: I0321 04:54:04.514076 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567814-fvs8k" Mar 21 04:54:04 crc kubenswrapper[4775]: I0321 04:54:04.530654 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-svf9w"] Mar 21 04:54:04 crc kubenswrapper[4775]: I0321 04:54:04.530870 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-svf9w" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" containerName="registry-server" containerID="cri-o://924b8554190993f0392bd8c20440e982df121b27ea4559dd16429c3faaecbb7a" gracePeriod=2 Mar 21 04:54:04 crc kubenswrapper[4775]: I0321 04:54:04.687309 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfgpl\" (UniqueName: \"kubernetes.io/projected/e8f9e3cb-7583-4349-98dc-422a7880fe5e-kube-api-access-gfgpl\") pod \"e8f9e3cb-7583-4349-98dc-422a7880fe5e\" (UID: \"e8f9e3cb-7583-4349-98dc-422a7880fe5e\") " Mar 21 04:54:04 crc kubenswrapper[4775]: I0321 04:54:04.696309 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8f9e3cb-7583-4349-98dc-422a7880fe5e-kube-api-access-gfgpl" (OuterVolumeSpecName: "kube-api-access-gfgpl") pod "e8f9e3cb-7583-4349-98dc-422a7880fe5e" (UID: "e8f9e3cb-7583-4349-98dc-422a7880fe5e"). InnerVolumeSpecName "kube-api-access-gfgpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:54:04 crc kubenswrapper[4775]: I0321 04:54:04.789679 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfgpl\" (UniqueName: \"kubernetes.io/projected/e8f9e3cb-7583-4349-98dc-422a7880fe5e-kube-api-access-gfgpl\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:04 crc kubenswrapper[4775]: I0321 04:54:04.963536 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svf9w" Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.092914 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dcbca72-150a-47c6-ac3c-f701ae82e05b-utilities\") pod \"1dcbca72-150a-47c6-ac3c-f701ae82e05b\" (UID: \"1dcbca72-150a-47c6-ac3c-f701ae82e05b\") " Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.093001 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dcbca72-150a-47c6-ac3c-f701ae82e05b-catalog-content\") pod \"1dcbca72-150a-47c6-ac3c-f701ae82e05b\" (UID: \"1dcbca72-150a-47c6-ac3c-f701ae82e05b\") " Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.093095 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wq48\" (UniqueName: \"kubernetes.io/projected/1dcbca72-150a-47c6-ac3c-f701ae82e05b-kube-api-access-4wq48\") pod \"1dcbca72-150a-47c6-ac3c-f701ae82e05b\" (UID: \"1dcbca72-150a-47c6-ac3c-f701ae82e05b\") " Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.093990 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dcbca72-150a-47c6-ac3c-f701ae82e05b-utilities" (OuterVolumeSpecName: "utilities") pod "1dcbca72-150a-47c6-ac3c-f701ae82e05b" (UID: "1dcbca72-150a-47c6-ac3c-f701ae82e05b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.096748 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dcbca72-150a-47c6-ac3c-f701ae82e05b-kube-api-access-4wq48" (OuterVolumeSpecName: "kube-api-access-4wq48") pod "1dcbca72-150a-47c6-ac3c-f701ae82e05b" (UID: "1dcbca72-150a-47c6-ac3c-f701ae82e05b"). InnerVolumeSpecName "kube-api-access-4wq48". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.143655 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dcbca72-150a-47c6-ac3c-f701ae82e05b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1dcbca72-150a-47c6-ac3c-f701ae82e05b" (UID: "1dcbca72-150a-47c6-ac3c-f701ae82e05b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.145111 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567814-fvs8k" event={"ID":"e8f9e3cb-7583-4349-98dc-422a7880fe5e","Type":"ContainerDied","Data":"bef13ceeba8d06c136483de90c1fc8ff328392147f003ed93dfc860c9a4863ee"} Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.145190 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567814-fvs8k" Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.145209 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bef13ceeba8d06c136483de90c1fc8ff328392147f003ed93dfc860c9a4863ee" Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.148084 4775 generic.go:334] "Generic (PLEG): container finished" podID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" containerID="924b8554190993f0392bd8c20440e982df121b27ea4559dd16429c3faaecbb7a" exitCode=0 Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.148272 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svf9w" event={"ID":"1dcbca72-150a-47c6-ac3c-f701ae82e05b","Type":"ContainerDied","Data":"924b8554190993f0392bd8c20440e982df121b27ea4559dd16429c3faaecbb7a"} Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.148313 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svf9w" event={"ID":"1dcbca72-150a-47c6-ac3c-f701ae82e05b","Type":"ContainerDied","Data":"e49106633d57c504f2a2b545d76a34559e4299f9b5c7c912c88ab2862f4e055a"} Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.148332 4775 scope.go:117] "RemoveContainer" containerID="924b8554190993f0392bd8c20440e982df121b27ea4559dd16429c3faaecbb7a" Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.148440 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svf9w" Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.163868 4775 scope.go:117] "RemoveContainer" containerID="a41e12fdcec460e076316274065e17d2ca1fc9dae7706fccd22acc387dfd57b1" Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.185350 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-svf9w"] Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.187191 4775 scope.go:117] "RemoveContainer" containerID="00e2e82d9b942bac6084d281d0485867290448469dd14201683721f46ea863fc" Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.193955 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wq48\" (UniqueName: \"kubernetes.io/projected/1dcbca72-150a-47c6-ac3c-f701ae82e05b-kube-api-access-4wq48\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.194021 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dcbca72-150a-47c6-ac3c-f701ae82e05b-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.194033 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dcbca72-150a-47c6-ac3c-f701ae82e05b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.198083 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-svf9w"] Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.201732 4775 scope.go:117] "RemoveContainer" containerID="924b8554190993f0392bd8c20440e982df121b27ea4559dd16429c3faaecbb7a" Mar 21 04:54:05 crc kubenswrapper[4775]: E0321 04:54:05.204795 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"924b8554190993f0392bd8c20440e982df121b27ea4559dd16429c3faaecbb7a\": container with ID starting with 924b8554190993f0392bd8c20440e982df121b27ea4559dd16429c3faaecbb7a not found: ID does not exist" containerID="924b8554190993f0392bd8c20440e982df121b27ea4559dd16429c3faaecbb7a" Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.204865 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"924b8554190993f0392bd8c20440e982df121b27ea4559dd16429c3faaecbb7a"} err="failed to get container status \"924b8554190993f0392bd8c20440e982df121b27ea4559dd16429c3faaecbb7a\": rpc error: code = NotFound desc = could not find container \"924b8554190993f0392bd8c20440e982df121b27ea4559dd16429c3faaecbb7a\": container with ID starting with 924b8554190993f0392bd8c20440e982df121b27ea4559dd16429c3faaecbb7a not found: ID does not exist" Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.204910 4775 scope.go:117] "RemoveContainer" containerID="a41e12fdcec460e076316274065e17d2ca1fc9dae7706fccd22acc387dfd57b1" Mar 21 04:54:05 crc kubenswrapper[4775]: E0321 04:54:05.205279 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a41e12fdcec460e076316274065e17d2ca1fc9dae7706fccd22acc387dfd57b1\": container with ID starting with a41e12fdcec460e076316274065e17d2ca1fc9dae7706fccd22acc387dfd57b1 not found: ID does not exist" containerID="a41e12fdcec460e076316274065e17d2ca1fc9dae7706fccd22acc387dfd57b1" Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.205311 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a41e12fdcec460e076316274065e17d2ca1fc9dae7706fccd22acc387dfd57b1"} err="failed to get container status \"a41e12fdcec460e076316274065e17d2ca1fc9dae7706fccd22acc387dfd57b1\": rpc error: code = NotFound desc = could not find container \"a41e12fdcec460e076316274065e17d2ca1fc9dae7706fccd22acc387dfd57b1\": container with ID starting with a41e12fdcec460e076316274065e17d2ca1fc9dae7706fccd22acc387dfd57b1 not found: ID does not exist" Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.205336 4775 scope.go:117] "RemoveContainer" containerID="00e2e82d9b942bac6084d281d0485867290448469dd14201683721f46ea863fc" Mar 21 04:54:05 crc kubenswrapper[4775]: E0321 04:54:05.205664 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00e2e82d9b942bac6084d281d0485867290448469dd14201683721f46ea863fc\": container with ID starting with 00e2e82d9b942bac6084d281d0485867290448469dd14201683721f46ea863fc not found: ID does not exist" containerID="00e2e82d9b942bac6084d281d0485867290448469dd14201683721f46ea863fc" Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.205698 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00e2e82d9b942bac6084d281d0485867290448469dd14201683721f46ea863fc"} err="failed to get container status \"00e2e82d9b942bac6084d281d0485867290448469dd14201683721f46ea863fc\": rpc error: code = NotFound desc = could not find container \"00e2e82d9b942bac6084d281d0485867290448469dd14201683721f46ea863fc\": container with ID starting with 00e2e82d9b942bac6084d281d0485867290448469dd14201683721f46ea863fc not found: ID does not exist" Mar 21 04:54:05 crc kubenswrapper[4775]: I0321 04:54:05.667182 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" path="/var/lib/kubelet/pods/1dcbca72-150a-47c6-ac3c-f701ae82e05b/volumes" Mar 21 04:54:06 crc kubenswrapper[4775]: I0321 04:54:06.731368 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7k9df"] Mar 21 04:54:06 crc kubenswrapper[4775]: I0321 04:54:06.731698 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7k9df" podUID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" containerName="registry-server" containerID="cri-o://e5e47a7de8c354ec132dd0116fce768c5a764f807e5342721397e5aed3144402" gracePeriod=2 Mar 21 04:54:06 crc kubenswrapper[4775]: I0321 04:54:06.881398 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4lpkr"] Mar 21 04:54:06 crc kubenswrapper[4775]: E0321 04:54:06.881662 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" containerName="extract-utilities" Mar 21 04:54:06 crc kubenswrapper[4775]: I0321 04:54:06.881679 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" containerName="extract-utilities" Mar 21 04:54:06 crc kubenswrapper[4775]: E0321 04:54:06.881695 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8f9e3cb-7583-4349-98dc-422a7880fe5e" containerName="oc" Mar 21 04:54:06 crc kubenswrapper[4775]: I0321 04:54:06.881702 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8f9e3cb-7583-4349-98dc-422a7880fe5e" containerName="oc" Mar 21 04:54:06 crc kubenswrapper[4775]: E0321 04:54:06.881712 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" containerName="registry-server" Mar 21 04:54:06 crc kubenswrapper[4775]: I0321 04:54:06.881719 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" containerName="registry-server" Mar 21 04:54:06 crc kubenswrapper[4775]: E0321 04:54:06.881732 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" containerName="extract-content" Mar 21 04:54:06 crc kubenswrapper[4775]: I0321 04:54:06.881740 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" containerName="extract-content" Mar 21 04:54:06 crc kubenswrapper[4775]: I0321 04:54:06.881860 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8f9e3cb-7583-4349-98dc-422a7880fe5e" containerName="oc" Mar 21 04:54:06 crc kubenswrapper[4775]: I0321 04:54:06.881876 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dcbca72-150a-47c6-ac3c-f701ae82e05b" containerName="registry-server" Mar 21 04:54:06 crc kubenswrapper[4775]: I0321 04:54:06.886047 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:06 crc kubenswrapper[4775]: I0321 04:54:06.904848 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4lpkr"] Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.015590 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/946ed25f-3a5b-446c-a928-2a85e719ff89-registry-tls\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.015644 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/946ed25f-3a5b-446c-a928-2a85e719ff89-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.015665 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/946ed25f-3a5b-446c-a928-2a85e719ff89-registry-certificates\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.015683 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/946ed25f-3a5b-446c-a928-2a85e719ff89-trusted-ca\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.015708 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/946ed25f-3a5b-446c-a928-2a85e719ff89-bound-sa-token\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.015796 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.015837 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pbvt\" (UniqueName: \"kubernetes.io/projected/946ed25f-3a5b-446c-a928-2a85e719ff89-kube-api-access-7pbvt\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.015867 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/946ed25f-3a5b-446c-a928-2a85e719ff89-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.039926 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.116455 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/946ed25f-3a5b-446c-a928-2a85e719ff89-registry-certificates\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.116512 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/946ed25f-3a5b-446c-a928-2a85e719ff89-trusted-ca\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.116540 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/946ed25f-3a5b-446c-a928-2a85e719ff89-bound-sa-token\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.116591 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pbvt\" (UniqueName: \"kubernetes.io/projected/946ed25f-3a5b-446c-a928-2a85e719ff89-kube-api-access-7pbvt\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.116621 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/946ed25f-3a5b-446c-a928-2a85e719ff89-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.116657 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/946ed25f-3a5b-446c-a928-2a85e719ff89-registry-tls\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.116694 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/946ed25f-3a5b-446c-a928-2a85e719ff89-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.117004 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/946ed25f-3a5b-446c-a928-2a85e719ff89-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.117697 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/946ed25f-3a5b-446c-a928-2a85e719ff89-registry-certificates\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.118672 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/946ed25f-3a5b-446c-a928-2a85e719ff89-trusted-ca\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.124951 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/946ed25f-3a5b-446c-a928-2a85e719ff89-registry-tls\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.129325 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/946ed25f-3a5b-446c-a928-2a85e719ff89-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.132951 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/946ed25f-3a5b-446c-a928-2a85e719ff89-bound-sa-token\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.133374 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pbvt\" (UniqueName: \"kubernetes.io/projected/946ed25f-3a5b-446c-a928-2a85e719ff89-kube-api-access-7pbvt\") pod \"image-registry-66df7c8f76-4lpkr\" (UID: \"946ed25f-3a5b-446c-a928-2a85e719ff89\") " pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: E0321 04:54:07.165029 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e5e47a7de8c354ec132dd0116fce768c5a764f807e5342721397e5aed3144402 is running failed: container process not found" containerID="e5e47a7de8c354ec132dd0116fce768c5a764f807e5342721397e5aed3144402" cmd=["grpc_health_probe","-addr=:50051"] Mar 21 04:54:07 crc kubenswrapper[4775]: E0321 04:54:07.165548 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e5e47a7de8c354ec132dd0116fce768c5a764f807e5342721397e5aed3144402 is running failed: container process not found" containerID="e5e47a7de8c354ec132dd0116fce768c5a764f807e5342721397e5aed3144402" cmd=["grpc_health_probe","-addr=:50051"] Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.165679 4775 generic.go:334] "Generic (PLEG): container finished" podID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" containerID="e5e47a7de8c354ec132dd0116fce768c5a764f807e5342721397e5aed3144402" exitCode=0 Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.165705 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7k9df" event={"ID":"55e2733e-3620-4ecb-a51e-33b1fd3dce9d","Type":"ContainerDied","Data":"e5e47a7de8c354ec132dd0116fce768c5a764f807e5342721397e5aed3144402"} Mar 21 04:54:07 crc kubenswrapper[4775]: E0321 04:54:07.165860 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e5e47a7de8c354ec132dd0116fce768c5a764f807e5342721397e5aed3144402 is running failed: container process not found" containerID="e5e47a7de8c354ec132dd0116fce768c5a764f807e5342721397e5aed3144402" cmd=["grpc_health_probe","-addr=:50051"] Mar 21 04:54:07 crc kubenswrapper[4775]: E0321 04:54:07.165892 4775 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e5e47a7de8c354ec132dd0116fce768c5a764f807e5342721397e5aed3144402 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-7k9df" podUID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" containerName="registry-server" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.192907 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7k9df" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.203541 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.318508 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55e2733e-3620-4ecb-a51e-33b1fd3dce9d-utilities\") pod \"55e2733e-3620-4ecb-a51e-33b1fd3dce9d\" (UID: \"55e2733e-3620-4ecb-a51e-33b1fd3dce9d\") " Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.319012 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mm7zf\" (UniqueName: \"kubernetes.io/projected/55e2733e-3620-4ecb-a51e-33b1fd3dce9d-kube-api-access-mm7zf\") pod \"55e2733e-3620-4ecb-a51e-33b1fd3dce9d\" (UID: \"55e2733e-3620-4ecb-a51e-33b1fd3dce9d\") " Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.319134 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55e2733e-3620-4ecb-a51e-33b1fd3dce9d-catalog-content\") pod \"55e2733e-3620-4ecb-a51e-33b1fd3dce9d\" (UID: \"55e2733e-3620-4ecb-a51e-33b1fd3dce9d\") " Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.320291 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55e2733e-3620-4ecb-a51e-33b1fd3dce9d-utilities" (OuterVolumeSpecName: "utilities") pod "55e2733e-3620-4ecb-a51e-33b1fd3dce9d" (UID: "55e2733e-3620-4ecb-a51e-33b1fd3dce9d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.322382 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55e2733e-3620-4ecb-a51e-33b1fd3dce9d-kube-api-access-mm7zf" (OuterVolumeSpecName: "kube-api-access-mm7zf") pod "55e2733e-3620-4ecb-a51e-33b1fd3dce9d" (UID: "55e2733e-3620-4ecb-a51e-33b1fd3dce9d"). InnerVolumeSpecName "kube-api-access-mm7zf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.324259 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55e2733e-3620-4ecb-a51e-33b1fd3dce9d-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.324307 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mm7zf\" (UniqueName: \"kubernetes.io/projected/55e2733e-3620-4ecb-a51e-33b1fd3dce9d-kube-api-access-mm7zf\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.361526 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55e2733e-3620-4ecb-a51e-33b1fd3dce9d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55e2733e-3620-4ecb-a51e-33b1fd3dce9d" (UID: "55e2733e-3620-4ecb-a51e-33b1fd3dce9d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.425405 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55e2733e-3620-4ecb-a51e-33b1fd3dce9d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:07 crc kubenswrapper[4775]: I0321 04:54:07.610749 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4lpkr"] Mar 21 04:54:07 crc kubenswrapper[4775]: W0321 04:54:07.615049 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod946ed25f_3a5b_446c_a928_2a85e719ff89.slice/crio-fa8267be73a89d641d14c54a74088f2114a7394e2ab393f72da796156c111a79 WatchSource:0}: Error finding container fa8267be73a89d641d14c54a74088f2114a7394e2ab393f72da796156c111a79: Status 404 returned error can't find the container with id fa8267be73a89d641d14c54a74088f2114a7394e2ab393f72da796156c111a79 Mar 21 04:54:08 crc kubenswrapper[4775]: I0321 04:54:08.171428 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" event={"ID":"946ed25f-3a5b-446c-a928-2a85e719ff89","Type":"ContainerStarted","Data":"89bc38e86d0a705f93814e892412fcdeffe60710cbc9847a1c18608f1666df9d"} Mar 21 04:54:08 crc kubenswrapper[4775]: I0321 04:54:08.171796 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" event={"ID":"946ed25f-3a5b-446c-a928-2a85e719ff89","Type":"ContainerStarted","Data":"fa8267be73a89d641d14c54a74088f2114a7394e2ab393f72da796156c111a79"} Mar 21 04:54:08 crc kubenswrapper[4775]: I0321 04:54:08.171842 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:08 crc kubenswrapper[4775]: I0321 04:54:08.173641 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7k9df" event={"ID":"55e2733e-3620-4ecb-a51e-33b1fd3dce9d","Type":"ContainerDied","Data":"5477663a3a0fc934d7191139776ed4c58017ab46346a0a37602d99108fd8b86f"} Mar 21 04:54:08 crc kubenswrapper[4775]: I0321 04:54:08.173692 4775 scope.go:117] "RemoveContainer" containerID="e5e47a7de8c354ec132dd0116fce768c5a764f807e5342721397e5aed3144402" Mar 21 04:54:08 crc kubenswrapper[4775]: I0321 04:54:08.173886 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7k9df" Mar 21 04:54:08 crc kubenswrapper[4775]: I0321 04:54:08.189078 4775 scope.go:117] "RemoveContainer" containerID="93e802bba3a0a60860befd9b343fb138a5913ba96b3cd2219f5f37eec14db557" Mar 21 04:54:08 crc kubenswrapper[4775]: I0321 04:54:08.195643 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" podStartSLOduration=2.195622179 podStartE2EDuration="2.195622179s" podCreationTimestamp="2026-03-21 04:54:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:54:08.190647187 +0000 UTC m=+401.167110821" watchObservedRunningTime="2026-03-21 04:54:08.195622179 +0000 UTC m=+401.172085803" Mar 21 04:54:08 crc kubenswrapper[4775]: I0321 04:54:08.202263 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7k9df"] Mar 21 04:54:08 crc kubenswrapper[4775]: I0321 04:54:08.206426 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7k9df"] Mar 21 04:54:08 crc kubenswrapper[4775]: I0321 04:54:08.209428 4775 scope.go:117] "RemoveContainer" containerID="2fca8a116d2bdda5fb8662833d28064b756cf76d00f4ea921c85b01d826cd2be" Mar 21 04:54:09 crc kubenswrapper[4775]: I0321 04:54:09.668798 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" path="/var/lib/kubelet/pods/55e2733e-3620-4ecb-a51e-33b1fd3dce9d/volumes" Mar 21 04:54:14 crc kubenswrapper[4775]: I0321 04:54:14.392275 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-77fc8489b9-p4xpt"] Mar 21 04:54:14 crc kubenswrapper[4775]: I0321 04:54:14.393081 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" podUID="41534621-ccd0-4e69-a13c-e6828feebe17" containerName="controller-manager" containerID="cri-o://5104548b211c2e16bc96a01a5e3e4f9c0385383b17796ee059444b9f78c92a94" gracePeriod=30 Mar 21 04:54:14 crc kubenswrapper[4775]: I0321 04:54:14.406771 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv"] Mar 21 04:54:14 crc kubenswrapper[4775]: I0321 04:54:14.407033 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" podUID="951da48c-671e-4004-85f0-2137949a3a2d" containerName="route-controller-manager" containerID="cri-o://3980fc5ac2c84abc585c64be5fe0e20cdc2ee02408d0873f172d592ea567b97a" gracePeriod=30 Mar 21 04:54:14 crc kubenswrapper[4775]: I0321 04:54:14.824536 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" Mar 21 04:54:14 crc kubenswrapper[4775]: I0321 04:54:14.919515 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4dwd\" (UniqueName: \"kubernetes.io/projected/951da48c-671e-4004-85f0-2137949a3a2d-kube-api-access-d4dwd\") pod \"951da48c-671e-4004-85f0-2137949a3a2d\" (UID: \"951da48c-671e-4004-85f0-2137949a3a2d\") " Mar 21 04:54:14 crc kubenswrapper[4775]: I0321 04:54:14.919564 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/951da48c-671e-4004-85f0-2137949a3a2d-config\") pod \"951da48c-671e-4004-85f0-2137949a3a2d\" (UID: \"951da48c-671e-4004-85f0-2137949a3a2d\") " Mar 21 04:54:14 crc kubenswrapper[4775]: I0321 04:54:14.919623 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/951da48c-671e-4004-85f0-2137949a3a2d-serving-cert\") pod \"951da48c-671e-4004-85f0-2137949a3a2d\" (UID: \"951da48c-671e-4004-85f0-2137949a3a2d\") " Mar 21 04:54:14 crc kubenswrapper[4775]: I0321 04:54:14.919676 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/951da48c-671e-4004-85f0-2137949a3a2d-client-ca\") pod \"951da48c-671e-4004-85f0-2137949a3a2d\" (UID: \"951da48c-671e-4004-85f0-2137949a3a2d\") " Mar 21 04:54:14 crc kubenswrapper[4775]: I0321 04:54:14.920610 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/951da48c-671e-4004-85f0-2137949a3a2d-config" (OuterVolumeSpecName: "config") pod "951da48c-671e-4004-85f0-2137949a3a2d" (UID: "951da48c-671e-4004-85f0-2137949a3a2d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:54:14 crc kubenswrapper[4775]: I0321 04:54:14.921577 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/951da48c-671e-4004-85f0-2137949a3a2d-client-ca" (OuterVolumeSpecName: "client-ca") pod "951da48c-671e-4004-85f0-2137949a3a2d" (UID: "951da48c-671e-4004-85f0-2137949a3a2d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:54:14 crc kubenswrapper[4775]: I0321 04:54:14.925359 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/951da48c-671e-4004-85f0-2137949a3a2d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "951da48c-671e-4004-85f0-2137949a3a2d" (UID: "951da48c-671e-4004-85f0-2137949a3a2d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:54:14 crc kubenswrapper[4775]: I0321 04:54:14.925559 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/951da48c-671e-4004-85f0-2137949a3a2d-kube-api-access-d4dwd" (OuterVolumeSpecName: "kube-api-access-d4dwd") pod "951da48c-671e-4004-85f0-2137949a3a2d" (UID: "951da48c-671e-4004-85f0-2137949a3a2d"). InnerVolumeSpecName "kube-api-access-d4dwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.009818 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.021040 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4dwd\" (UniqueName: \"kubernetes.io/projected/951da48c-671e-4004-85f0-2137949a3a2d-kube-api-access-d4dwd\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.021096 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/951da48c-671e-4004-85f0-2137949a3a2d-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.021106 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/951da48c-671e-4004-85f0-2137949a3a2d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.021129 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/951da48c-671e-4004-85f0-2137949a3a2d-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.121674 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41534621-ccd0-4e69-a13c-e6828feebe17-config\") pod \"41534621-ccd0-4e69-a13c-e6828feebe17\" (UID: \"41534621-ccd0-4e69-a13c-e6828feebe17\") " Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.121750 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41534621-ccd0-4e69-a13c-e6828feebe17-serving-cert\") pod \"41534621-ccd0-4e69-a13c-e6828feebe17\" (UID: \"41534621-ccd0-4e69-a13c-e6828feebe17\") " Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.121834 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/41534621-ccd0-4e69-a13c-e6828feebe17-proxy-ca-bundles\") pod \"41534621-ccd0-4e69-a13c-e6828feebe17\" (UID: \"41534621-ccd0-4e69-a13c-e6828feebe17\") " Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.121907 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s48s5\" (UniqueName: \"kubernetes.io/projected/41534621-ccd0-4e69-a13c-e6828feebe17-kube-api-access-s48s5\") pod \"41534621-ccd0-4e69-a13c-e6828feebe17\" (UID: \"41534621-ccd0-4e69-a13c-e6828feebe17\") " Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.122045 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41534621-ccd0-4e69-a13c-e6828feebe17-client-ca\") pod \"41534621-ccd0-4e69-a13c-e6828feebe17\" (UID: \"41534621-ccd0-4e69-a13c-e6828feebe17\") " Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.122689 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41534621-ccd0-4e69-a13c-e6828feebe17-config" (OuterVolumeSpecName: "config") pod "41534621-ccd0-4e69-a13c-e6828feebe17" (UID: "41534621-ccd0-4e69-a13c-e6828feebe17"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.122710 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41534621-ccd0-4e69-a13c-e6828feebe17-client-ca" (OuterVolumeSpecName: "client-ca") pod "41534621-ccd0-4e69-a13c-e6828feebe17" (UID: "41534621-ccd0-4e69-a13c-e6828feebe17"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.123425 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41534621-ccd0-4e69-a13c-e6828feebe17-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "41534621-ccd0-4e69-a13c-e6828feebe17" (UID: "41534621-ccd0-4e69-a13c-e6828feebe17"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.125714 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41534621-ccd0-4e69-a13c-e6828feebe17-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "41534621-ccd0-4e69-a13c-e6828feebe17" (UID: "41534621-ccd0-4e69-a13c-e6828feebe17"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.126862 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41534621-ccd0-4e69-a13c-e6828feebe17-kube-api-access-s48s5" (OuterVolumeSpecName: "kube-api-access-s48s5") pod "41534621-ccd0-4e69-a13c-e6828feebe17" (UID: "41534621-ccd0-4e69-a13c-e6828feebe17"). InnerVolumeSpecName "kube-api-access-s48s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.208442 4775 generic.go:334] "Generic (PLEG): container finished" podID="951da48c-671e-4004-85f0-2137949a3a2d" containerID="3980fc5ac2c84abc585c64be5fe0e20cdc2ee02408d0873f172d592ea567b97a" exitCode=0 Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.208487 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" event={"ID":"951da48c-671e-4004-85f0-2137949a3a2d","Type":"ContainerDied","Data":"3980fc5ac2c84abc585c64be5fe0e20cdc2ee02408d0873f172d592ea567b97a"} Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.208503 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.208537 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv" event={"ID":"951da48c-671e-4004-85f0-2137949a3a2d","Type":"ContainerDied","Data":"aa646cd6ba5a44de57d221735dcb1de71958eda8b620088cc7ef6124aed0413e"} Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.208562 4775 scope.go:117] "RemoveContainer" containerID="3980fc5ac2c84abc585c64be5fe0e20cdc2ee02408d0873f172d592ea567b97a" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.210440 4775 generic.go:334] "Generic (PLEG): container finished" podID="41534621-ccd0-4e69-a13c-e6828feebe17" containerID="5104548b211c2e16bc96a01a5e3e4f9c0385383b17796ee059444b9f78c92a94" exitCode=0 Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.210474 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" event={"ID":"41534621-ccd0-4e69-a13c-e6828feebe17","Type":"ContainerDied","Data":"5104548b211c2e16bc96a01a5e3e4f9c0385383b17796ee059444b9f78c92a94"} Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.210490 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" event={"ID":"41534621-ccd0-4e69-a13c-e6828feebe17","Type":"ContainerDied","Data":"69122f1ec9bc520b92bd94027f2fae1d2dd0387607089568975168c52c43b3a7"} Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.210541 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77fc8489b9-p4xpt" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.223912 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s48s5\" (UniqueName: \"kubernetes.io/projected/41534621-ccd0-4e69-a13c-e6828feebe17-kube-api-access-s48s5\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.223950 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41534621-ccd0-4e69-a13c-e6828feebe17-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.223962 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41534621-ccd0-4e69-a13c-e6828feebe17-config\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.223972 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41534621-ccd0-4e69-a13c-e6828feebe17-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.223983 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/41534621-ccd0-4e69-a13c-e6828feebe17-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.225546 4775 scope.go:117] "RemoveContainer" containerID="3980fc5ac2c84abc585c64be5fe0e20cdc2ee02408d0873f172d592ea567b97a" Mar 21 04:54:15 crc kubenswrapper[4775]: E0321 04:54:15.226080 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3980fc5ac2c84abc585c64be5fe0e20cdc2ee02408d0873f172d592ea567b97a\": container with ID starting with 3980fc5ac2c84abc585c64be5fe0e20cdc2ee02408d0873f172d592ea567b97a not found: ID does not exist" containerID="3980fc5ac2c84abc585c64be5fe0e20cdc2ee02408d0873f172d592ea567b97a" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.226244 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3980fc5ac2c84abc585c64be5fe0e20cdc2ee02408d0873f172d592ea567b97a"} err="failed to get container status \"3980fc5ac2c84abc585c64be5fe0e20cdc2ee02408d0873f172d592ea567b97a\": rpc error: code = NotFound desc = could not find container \"3980fc5ac2c84abc585c64be5fe0e20cdc2ee02408d0873f172d592ea567b97a\": container with ID starting with 3980fc5ac2c84abc585c64be5fe0e20cdc2ee02408d0873f172d592ea567b97a not found: ID does not exist" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.226305 4775 scope.go:117] "RemoveContainer" containerID="5104548b211c2e16bc96a01a5e3e4f9c0385383b17796ee059444b9f78c92a94" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.255549 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-77fc8489b9-p4xpt"] Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.258916 4775 scope.go:117] "RemoveContainer" containerID="5104548b211c2e16bc96a01a5e3e4f9c0385383b17796ee059444b9f78c92a94" Mar 21 04:54:15 crc kubenswrapper[4775]: E0321 04:54:15.259798 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5104548b211c2e16bc96a01a5e3e4f9c0385383b17796ee059444b9f78c92a94\": container with ID starting with 5104548b211c2e16bc96a01a5e3e4f9c0385383b17796ee059444b9f78c92a94 not found: ID does not exist" containerID="5104548b211c2e16bc96a01a5e3e4f9c0385383b17796ee059444b9f78c92a94" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.259833 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5104548b211c2e16bc96a01a5e3e4f9c0385383b17796ee059444b9f78c92a94"} err="failed to get container status \"5104548b211c2e16bc96a01a5e3e4f9c0385383b17796ee059444b9f78c92a94\": rpc error: code = NotFound desc = could not find container \"5104548b211c2e16bc96a01a5e3e4f9c0385383b17796ee059444b9f78c92a94\": container with ID starting with 5104548b211c2e16bc96a01a5e3e4f9c0385383b17796ee059444b9f78c92a94 not found: ID does not exist" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.266321 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-77fc8489b9-p4xpt"] Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.271833 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv"] Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.276288 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d78b4c7b9-244bv"] Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.668835 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41534621-ccd0-4e69-a13c-e6828feebe17" path="/var/lib/kubelet/pods/41534621-ccd0-4e69-a13c-e6828feebe17/volumes" Mar 21 04:54:15 crc kubenswrapper[4775]: I0321 04:54:15.669360 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="951da48c-671e-4004-85f0-2137949a3a2d" path="/var/lib/kubelet/pods/951da48c-671e-4004-85f0-2137949a3a2d/volumes" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.047800 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh"] Mar 21 04:54:16 crc kubenswrapper[4775]: E0321 04:54:16.048081 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" containerName="extract-content" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.048146 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" containerName="extract-content" Mar 21 04:54:16 crc kubenswrapper[4775]: E0321 04:54:16.048158 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41534621-ccd0-4e69-a13c-e6828feebe17" containerName="controller-manager" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.048165 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="41534621-ccd0-4e69-a13c-e6828feebe17" containerName="controller-manager" Mar 21 04:54:16 crc kubenswrapper[4775]: E0321 04:54:16.048177 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" containerName="extract-utilities" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.048184 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" containerName="extract-utilities" Mar 21 04:54:16 crc kubenswrapper[4775]: E0321 04:54:16.048198 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="951da48c-671e-4004-85f0-2137949a3a2d" containerName="route-controller-manager" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.048204 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="951da48c-671e-4004-85f0-2137949a3a2d" containerName="route-controller-manager" Mar 21 04:54:16 crc kubenswrapper[4775]: E0321 04:54:16.048219 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" containerName="registry-server" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.048226 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" containerName="registry-server" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.048346 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="55e2733e-3620-4ecb-a51e-33b1fd3dce9d" containerName="registry-server" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.048361 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="41534621-ccd0-4e69-a13c-e6828feebe17" containerName="controller-manager" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.048370 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="951da48c-671e-4004-85f0-2137949a3a2d" containerName="route-controller-manager" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.048840 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.051298 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv"] Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.051894 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.057565 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.073137 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.073184 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.073210 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.073342 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.073461 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.073922 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.074057 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.074199 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.074238 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.074460 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.077195 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.077895 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh"] Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.085378 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.091715 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv"] Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.135388 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0aff99b-e78c-42b5-ac13-8fb27b6c88e8-serving-cert\") pod \"controller-manager-5b5b8879f4-p7rtv\" (UID: \"e0aff99b-e78c-42b5-ac13-8fb27b6c88e8\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.135451 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8b446038-617c-4df4-8bbe-5d821cd3dc27-client-ca\") pod \"route-controller-manager-555c6b4d6b-pkcfh\" (UID: \"8b446038-617c-4df4-8bbe-5d821cd3dc27\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.135488 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ngsr\" (UniqueName: \"kubernetes.io/projected/8b446038-617c-4df4-8bbe-5d821cd3dc27-kube-api-access-6ngsr\") pod \"route-controller-manager-555c6b4d6b-pkcfh\" (UID: \"8b446038-617c-4df4-8bbe-5d821cd3dc27\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.135531 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e0aff99b-e78c-42b5-ac13-8fb27b6c88e8-proxy-ca-bundles\") pod \"controller-manager-5b5b8879f4-p7rtv\" (UID: \"e0aff99b-e78c-42b5-ac13-8fb27b6c88e8\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.135569 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0aff99b-e78c-42b5-ac13-8fb27b6c88e8-config\") pod \"controller-manager-5b5b8879f4-p7rtv\" (UID: \"e0aff99b-e78c-42b5-ac13-8fb27b6c88e8\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.135596 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0aff99b-e78c-42b5-ac13-8fb27b6c88e8-client-ca\") pod \"controller-manager-5b5b8879f4-p7rtv\" (UID: \"e0aff99b-e78c-42b5-ac13-8fb27b6c88e8\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.135618 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b446038-617c-4df4-8bbe-5d821cd3dc27-config\") pod \"route-controller-manager-555c6b4d6b-pkcfh\" (UID: \"8b446038-617c-4df4-8bbe-5d821cd3dc27\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.135639 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khn9s\" (UniqueName: \"kubernetes.io/projected/e0aff99b-e78c-42b5-ac13-8fb27b6c88e8-kube-api-access-khn9s\") pod \"controller-manager-5b5b8879f4-p7rtv\" (UID: \"e0aff99b-e78c-42b5-ac13-8fb27b6c88e8\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.135665 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b446038-617c-4df4-8bbe-5d821cd3dc27-serving-cert\") pod \"route-controller-manager-555c6b4d6b-pkcfh\" (UID: \"8b446038-617c-4df4-8bbe-5d821cd3dc27\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.236357 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0aff99b-e78c-42b5-ac13-8fb27b6c88e8-config\") pod \"controller-manager-5b5b8879f4-p7rtv\" (UID: \"e0aff99b-e78c-42b5-ac13-8fb27b6c88e8\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.236411 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0aff99b-e78c-42b5-ac13-8fb27b6c88e8-client-ca\") pod \"controller-manager-5b5b8879f4-p7rtv\" (UID: \"e0aff99b-e78c-42b5-ac13-8fb27b6c88e8\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.236442 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b446038-617c-4df4-8bbe-5d821cd3dc27-config\") pod \"route-controller-manager-555c6b4d6b-pkcfh\" (UID: \"8b446038-617c-4df4-8bbe-5d821cd3dc27\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.236462 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khn9s\" (UniqueName: \"kubernetes.io/projected/e0aff99b-e78c-42b5-ac13-8fb27b6c88e8-kube-api-access-khn9s\") pod \"controller-manager-5b5b8879f4-p7rtv\" (UID: \"e0aff99b-e78c-42b5-ac13-8fb27b6c88e8\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.236484 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b446038-617c-4df4-8bbe-5d821cd3dc27-serving-cert\") pod \"route-controller-manager-555c6b4d6b-pkcfh\" (UID: \"8b446038-617c-4df4-8bbe-5d821cd3dc27\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.236502 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0aff99b-e78c-42b5-ac13-8fb27b6c88e8-serving-cert\") pod \"controller-manager-5b5b8879f4-p7rtv\" (UID: \"e0aff99b-e78c-42b5-ac13-8fb27b6c88e8\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.236524 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8b446038-617c-4df4-8bbe-5d821cd3dc27-client-ca\") pod \"route-controller-manager-555c6b4d6b-pkcfh\" (UID: \"8b446038-617c-4df4-8bbe-5d821cd3dc27\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.236549 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ngsr\" (UniqueName: \"kubernetes.io/projected/8b446038-617c-4df4-8bbe-5d821cd3dc27-kube-api-access-6ngsr\") pod \"route-controller-manager-555c6b4d6b-pkcfh\" (UID: \"8b446038-617c-4df4-8bbe-5d821cd3dc27\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.236582 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e0aff99b-e78c-42b5-ac13-8fb27b6c88e8-proxy-ca-bundles\") pod \"controller-manager-5b5b8879f4-p7rtv\" (UID: \"e0aff99b-e78c-42b5-ac13-8fb27b6c88e8\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.237424 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0aff99b-e78c-42b5-ac13-8fb27b6c88e8-client-ca\") pod \"controller-manager-5b5b8879f4-p7rtv\" (UID: \"e0aff99b-e78c-42b5-ac13-8fb27b6c88e8\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.237658 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0aff99b-e78c-42b5-ac13-8fb27b6c88e8-config\") pod \"controller-manager-5b5b8879f4-p7rtv\" (UID: \"e0aff99b-e78c-42b5-ac13-8fb27b6c88e8\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.237777 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e0aff99b-e78c-42b5-ac13-8fb27b6c88e8-proxy-ca-bundles\") pod \"controller-manager-5b5b8879f4-p7rtv\" (UID: \"e0aff99b-e78c-42b5-ac13-8fb27b6c88e8\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.237792 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b446038-617c-4df4-8bbe-5d821cd3dc27-config\") pod \"route-controller-manager-555c6b4d6b-pkcfh\" (UID: \"8b446038-617c-4df4-8bbe-5d821cd3dc27\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.238265 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8b446038-617c-4df4-8bbe-5d821cd3dc27-client-ca\") pod \"route-controller-manager-555c6b4d6b-pkcfh\" (UID: \"8b446038-617c-4df4-8bbe-5d821cd3dc27\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.240783 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0aff99b-e78c-42b5-ac13-8fb27b6c88e8-serving-cert\") pod \"controller-manager-5b5b8879f4-p7rtv\" (UID: \"e0aff99b-e78c-42b5-ac13-8fb27b6c88e8\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.242105 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b446038-617c-4df4-8bbe-5d821cd3dc27-serving-cert\") pod \"route-controller-manager-555c6b4d6b-pkcfh\" (UID: \"8b446038-617c-4df4-8bbe-5d821cd3dc27\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.256433 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khn9s\" (UniqueName: \"kubernetes.io/projected/e0aff99b-e78c-42b5-ac13-8fb27b6c88e8-kube-api-access-khn9s\") pod \"controller-manager-5b5b8879f4-p7rtv\" (UID: \"e0aff99b-e78c-42b5-ac13-8fb27b6c88e8\") " pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.261836 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ngsr\" (UniqueName: \"kubernetes.io/projected/8b446038-617c-4df4-8bbe-5d821cd3dc27-kube-api-access-6ngsr\") pod \"route-controller-manager-555c6b4d6b-pkcfh\" (UID: \"8b446038-617c-4df4-8bbe-5d821cd3dc27\") " pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.365640 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.377078 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.693838 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh"] Mar 21 04:54:16 crc kubenswrapper[4775]: I0321 04:54:16.833442 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv"] Mar 21 04:54:16 crc kubenswrapper[4775]: W0321 04:54:16.851676 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0aff99b_e78c_42b5_ac13_8fb27b6c88e8.slice/crio-1fe730cb1afcc8a1047dbfae63ddc7b521943634d7feffb06f13e3de2a8a64cb WatchSource:0}: Error finding container 1fe730cb1afcc8a1047dbfae63ddc7b521943634d7feffb06f13e3de2a8a64cb: Status 404 returned error can't find the container with id 1fe730cb1afcc8a1047dbfae63ddc7b521943634d7feffb06f13e3de2a8a64cb Mar 21 04:54:17 crc kubenswrapper[4775]: I0321 04:54:17.223924 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh" event={"ID":"8b446038-617c-4df4-8bbe-5d821cd3dc27","Type":"ContainerStarted","Data":"4aaf40b75759912b5aef27b45aeb115f7739c6aab50515e03ff4a232d96bd5c7"} Mar 21 04:54:17 crc kubenswrapper[4775]: I0321 04:54:17.223977 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh" event={"ID":"8b446038-617c-4df4-8bbe-5d821cd3dc27","Type":"ContainerStarted","Data":"3a7d4dcc62dee004c6591691be664dbc82a6d433863cc1859414b3800c88b76f"} Mar 21 04:54:17 crc kubenswrapper[4775]: I0321 04:54:17.225136 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh" Mar 21 04:54:17 crc kubenswrapper[4775]: I0321 04:54:17.226606 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" event={"ID":"e0aff99b-e78c-42b5-ac13-8fb27b6c88e8","Type":"ContainerStarted","Data":"1c522f7b9a2336087d2eee63f9397f360e81837d6c247583783be9f284c41a0f"} Mar 21 04:54:17 crc kubenswrapper[4775]: I0321 04:54:17.226643 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" event={"ID":"e0aff99b-e78c-42b5-ac13-8fb27b6c88e8","Type":"ContainerStarted","Data":"1fe730cb1afcc8a1047dbfae63ddc7b521943634d7feffb06f13e3de2a8a64cb"} Mar 21 04:54:17 crc kubenswrapper[4775]: I0321 04:54:17.226839 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" Mar 21 04:54:17 crc kubenswrapper[4775]: I0321 04:54:17.231518 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" Mar 21 04:54:17 crc kubenswrapper[4775]: I0321 04:54:17.250411 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh" podStartSLOduration=3.250392365 podStartE2EDuration="3.250392365s" podCreationTimestamp="2026-03-21 04:54:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:54:17.246193404 +0000 UTC m=+410.222657038" watchObservedRunningTime="2026-03-21 04:54:17.250392365 +0000 UTC m=+410.226855979" Mar 21 04:54:17 crc kubenswrapper[4775]: I0321 04:54:17.271810 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5b5b8879f4-p7rtv" podStartSLOduration=3.2717912350000002 podStartE2EDuration="3.271791235s" podCreationTimestamp="2026-03-21 04:54:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:54:17.271190459 +0000 UTC m=+410.247654093" watchObservedRunningTime="2026-03-21 04:54:17.271791235 +0000 UTC m=+410.248254859" Mar 21 04:54:17 crc kubenswrapper[4775]: I0321 04:54:17.277012 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-555c6b4d6b-pkcfh" Mar 21 04:54:27 crc kubenswrapper[4775]: I0321 04:54:27.210826 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-4lpkr" Mar 21 04:54:27 crc kubenswrapper[4775]: I0321 04:54:27.271066 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hvd8t"] Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.425578 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-959kj"] Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.430039 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-959kj" podUID="5bb65dee-cd5f-46b3-9e7d-36e5d182d19e" containerName="registry-server" containerID="cri-o://aaa815a5f0e220d5cd57bb1c4b415350dcf17e31c1b987c38d4e0714c4280f96" gracePeriod=30 Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.435172 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wbtgv"] Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.443710 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wbtgv" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" containerName="registry-server" containerID="cri-o://02ab090064cf43714a361f98c8401a08ff5684170017c160d3affd080275fc5a" gracePeriod=30 Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.444212 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wrzs4"] Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.444409 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" podUID="dfbaac71-f99c-4373-a469-f2e5dd0ee632" containerName="marketplace-operator" containerID="cri-o://53e93eb8775f797ec5701d0054608e0cc06d124aa1940ad067f4c33b12a82fa2" gracePeriod=30 Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.456327 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hldm7"] Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.456609 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hldm7" podUID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" containerName="registry-server" containerID="cri-o://bce5cf8972de39d9de712cead6829664ca56252c15f4837eb08ebb6d9c3a039b" gracePeriod=30 Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.470912 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z7fcx"] Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.471814 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z7fcx" Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.483533 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wjfwd"] Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.483893 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wjfwd" podUID="70ad413a-5f81-4094-b2d8-9b89698c6e32" containerName="registry-server" containerID="cri-o://dd9cc974af699ec57297810c7016806ecad201037353fbcf0dc303241f78fc0b" gracePeriod=30 Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.485011 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z7fcx"] Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.632019 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmbw8\" (UniqueName: \"kubernetes.io/projected/59fec450-4b61-4a15-b1b5-b47dedd649a0-kube-api-access-zmbw8\") pod \"marketplace-operator-79b997595-z7fcx\" (UID: \"59fec450-4b61-4a15-b1b5-b47dedd649a0\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7fcx" Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.632095 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/59fec450-4b61-4a15-b1b5-b47dedd649a0-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z7fcx\" (UID: \"59fec450-4b61-4a15-b1b5-b47dedd649a0\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7fcx" Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.632248 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/59fec450-4b61-4a15-b1b5-b47dedd649a0-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z7fcx\" (UID: \"59fec450-4b61-4a15-b1b5-b47dedd649a0\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7fcx" Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.734082 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmbw8\" (UniqueName: \"kubernetes.io/projected/59fec450-4b61-4a15-b1b5-b47dedd649a0-kube-api-access-zmbw8\") pod \"marketplace-operator-79b997595-z7fcx\" (UID: \"59fec450-4b61-4a15-b1b5-b47dedd649a0\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7fcx" Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.734170 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/59fec450-4b61-4a15-b1b5-b47dedd649a0-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z7fcx\" (UID: \"59fec450-4b61-4a15-b1b5-b47dedd649a0\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7fcx" Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.734229 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/59fec450-4b61-4a15-b1b5-b47dedd649a0-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z7fcx\" (UID: \"59fec450-4b61-4a15-b1b5-b47dedd649a0\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7fcx" Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.735769 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/59fec450-4b61-4a15-b1b5-b47dedd649a0-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z7fcx\" (UID: \"59fec450-4b61-4a15-b1b5-b47dedd649a0\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7fcx" Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.741771 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/59fec450-4b61-4a15-b1b5-b47dedd649a0-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z7fcx\" (UID: \"59fec450-4b61-4a15-b1b5-b47dedd649a0\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7fcx" Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.750919 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmbw8\" (UniqueName: \"kubernetes.io/projected/59fec450-4b61-4a15-b1b5-b47dedd649a0-kube-api-access-zmbw8\") pod \"marketplace-operator-79b997595-z7fcx\" (UID: \"59fec450-4b61-4a15-b1b5-b47dedd649a0\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7fcx" Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.800391 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z7fcx" Mar 21 04:54:28 crc kubenswrapper[4775]: I0321 04:54:28.987344 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wbtgv" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.139498 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccb910ab-dcef-4523-81df-c0fb5eb83429-catalog-content\") pod \"ccb910ab-dcef-4523-81df-c0fb5eb83429\" (UID: \"ccb910ab-dcef-4523-81df-c0fb5eb83429\") " Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.139567 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccb910ab-dcef-4523-81df-c0fb5eb83429-utilities\") pod \"ccb910ab-dcef-4523-81df-c0fb5eb83429\" (UID: \"ccb910ab-dcef-4523-81df-c0fb5eb83429\") " Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.139596 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92fzw\" (UniqueName: \"kubernetes.io/projected/ccb910ab-dcef-4523-81df-c0fb5eb83429-kube-api-access-92fzw\") pod \"ccb910ab-dcef-4523-81df-c0fb5eb83429\" (UID: \"ccb910ab-dcef-4523-81df-c0fb5eb83429\") " Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.140746 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccb910ab-dcef-4523-81df-c0fb5eb83429-utilities" (OuterVolumeSpecName: "utilities") pod "ccb910ab-dcef-4523-81df-c0fb5eb83429" (UID: "ccb910ab-dcef-4523-81df-c0fb5eb83429"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.144271 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccb910ab-dcef-4523-81df-c0fb5eb83429-kube-api-access-92fzw" (OuterVolumeSpecName: "kube-api-access-92fzw") pod "ccb910ab-dcef-4523-81df-c0fb5eb83429" (UID: "ccb910ab-dcef-4523-81df-c0fb5eb83429"). InnerVolumeSpecName "kube-api-access-92fzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.176575 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-959kj" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.183583 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wjfwd" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.185299 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hldm7" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.215133 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccb910ab-dcef-4523-81df-c0fb5eb83429-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ccb910ab-dcef-4523-81df-c0fb5eb83429" (UID: "ccb910ab-dcef-4523-81df-c0fb5eb83429"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.240552 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccb910ab-dcef-4523-81df-c0fb5eb83429-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.240585 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccb910ab-dcef-4523-81df-c0fb5eb83429-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.240603 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92fzw\" (UniqueName: \"kubernetes.io/projected/ccb910ab-dcef-4523-81df-c0fb5eb83429-kube-api-access-92fzw\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.241434 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.313950 4775 generic.go:334] "Generic (PLEG): container finished" podID="70ad413a-5f81-4094-b2d8-9b89698c6e32" containerID="dd9cc974af699ec57297810c7016806ecad201037353fbcf0dc303241f78fc0b" exitCode=0 Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.314024 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjfwd" event={"ID":"70ad413a-5f81-4094-b2d8-9b89698c6e32","Type":"ContainerDied","Data":"dd9cc974af699ec57297810c7016806ecad201037353fbcf0dc303241f78fc0b"} Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.314051 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjfwd" event={"ID":"70ad413a-5f81-4094-b2d8-9b89698c6e32","Type":"ContainerDied","Data":"f3ec71e4b8bce118032bd210cfae1bbf7c53e6c8a5aba3f4f045bc35e188b3c0"} Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.314067 4775 scope.go:117] "RemoveContainer" containerID="dd9cc974af699ec57297810c7016806ecad201037353fbcf0dc303241f78fc0b" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.314218 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wjfwd" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.318732 4775 generic.go:334] "Generic (PLEG): container finished" podID="5bb65dee-cd5f-46b3-9e7d-36e5d182d19e" containerID="aaa815a5f0e220d5cd57bb1c4b415350dcf17e31c1b987c38d4e0714c4280f96" exitCode=0 Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.318806 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-959kj" event={"ID":"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e","Type":"ContainerDied","Data":"aaa815a5f0e220d5cd57bb1c4b415350dcf17e31c1b987c38d4e0714c4280f96"} Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.318849 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-959kj" event={"ID":"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e","Type":"ContainerDied","Data":"68f05861e537811c2b45eab073d2762025cd79d47c769cec2072bbee0ee4a322"} Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.318929 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-959kj" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.322498 4775 generic.go:334] "Generic (PLEG): container finished" podID="dfbaac71-f99c-4373-a469-f2e5dd0ee632" containerID="53e93eb8775f797ec5701d0054608e0cc06d124aa1940ad067f4c33b12a82fa2" exitCode=0 Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.322585 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" event={"ID":"dfbaac71-f99c-4373-a469-f2e5dd0ee632","Type":"ContainerDied","Data":"53e93eb8775f797ec5701d0054608e0cc06d124aa1940ad067f4c33b12a82fa2"} Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.322621 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" event={"ID":"dfbaac71-f99c-4373-a469-f2e5dd0ee632","Type":"ContainerDied","Data":"73bfd1fd04d6bcc9e68b01d3bd224e4608501bca9349188673dd3d6681965588"} Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.322692 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wrzs4" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.326637 4775 generic.go:334] "Generic (PLEG): container finished" podID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" containerID="bce5cf8972de39d9de712cead6829664ca56252c15f4837eb08ebb6d9c3a039b" exitCode=0 Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.326702 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hldm7" event={"ID":"571e84f2-a2bc-4f09-ac53-d4a4adafa80b","Type":"ContainerDied","Data":"bce5cf8972de39d9de712cead6829664ca56252c15f4837eb08ebb6d9c3a039b"} Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.326732 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hldm7" event={"ID":"571e84f2-a2bc-4f09-ac53-d4a4adafa80b","Type":"ContainerDied","Data":"615ed550bf3a5c04a2f6727d10336a359668c9d0754381b6a62de7c552e1b1b1"} Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.327009 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hldm7" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.329210 4775 generic.go:334] "Generic (PLEG): container finished" podID="ccb910ab-dcef-4523-81df-c0fb5eb83429" containerID="02ab090064cf43714a361f98c8401a08ff5684170017c160d3affd080275fc5a" exitCode=0 Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.329248 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbtgv" event={"ID":"ccb910ab-dcef-4523-81df-c0fb5eb83429","Type":"ContainerDied","Data":"02ab090064cf43714a361f98c8401a08ff5684170017c160d3affd080275fc5a"} Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.329269 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wbtgv" event={"ID":"ccb910ab-dcef-4523-81df-c0fb5eb83429","Type":"ContainerDied","Data":"c9000a4269c650286aea4fb24ec488fdb1ea75dd0febdc2786cb6046e9da1f69"} Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.329342 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wbtgv" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.333435 4775 scope.go:117] "RemoveContainer" containerID="1098f33834501f7272dae8532213ab991de93513a5cd5e8028551095da4598b0" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.380334 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bb65dee-cd5f-46b3-9e7d-36e5d182d19e-utilities\") pod \"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e\" (UID: \"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e\") " Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.380381 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpppf\" (UniqueName: \"kubernetes.io/projected/70ad413a-5f81-4094-b2d8-9b89698c6e32-kube-api-access-vpppf\") pod \"70ad413a-5f81-4094-b2d8-9b89698c6e32\" (UID: \"70ad413a-5f81-4094-b2d8-9b89698c6e32\") " Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.380426 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dfbaac71-f99c-4373-a469-f2e5dd0ee632-marketplace-operator-metrics\") pod \"dfbaac71-f99c-4373-a469-f2e5dd0ee632\" (UID: \"dfbaac71-f99c-4373-a469-f2e5dd0ee632\") " Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.380446 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m44cw\" (UniqueName: \"kubernetes.io/projected/dfbaac71-f99c-4373-a469-f2e5dd0ee632-kube-api-access-m44cw\") pod \"dfbaac71-f99c-4373-a469-f2e5dd0ee632\" (UID: \"dfbaac71-f99c-4373-a469-f2e5dd0ee632\") " Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.380476 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/571e84f2-a2bc-4f09-ac53-d4a4adafa80b-catalog-content\") pod \"571e84f2-a2bc-4f09-ac53-d4a4adafa80b\" (UID: \"571e84f2-a2bc-4f09-ac53-d4a4adafa80b\") " Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.380497 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bb65dee-cd5f-46b3-9e7d-36e5d182d19e-catalog-content\") pod \"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e\" (UID: \"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e\") " Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.380521 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70ad413a-5f81-4094-b2d8-9b89698c6e32-utilities\") pod \"70ad413a-5f81-4094-b2d8-9b89698c6e32\" (UID: \"70ad413a-5f81-4094-b2d8-9b89698c6e32\") " Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.380550 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/571e84f2-a2bc-4f09-ac53-d4a4adafa80b-utilities\") pod \"571e84f2-a2bc-4f09-ac53-d4a4adafa80b\" (UID: \"571e84f2-a2bc-4f09-ac53-d4a4adafa80b\") " Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.380567 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ww6h\" (UniqueName: \"kubernetes.io/projected/5bb65dee-cd5f-46b3-9e7d-36e5d182d19e-kube-api-access-7ww6h\") pod \"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e\" (UID: \"5bb65dee-cd5f-46b3-9e7d-36e5d182d19e\") " Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.380596 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70ad413a-5f81-4094-b2d8-9b89698c6e32-catalog-content\") pod \"70ad413a-5f81-4094-b2d8-9b89698c6e32\" (UID: \"70ad413a-5f81-4094-b2d8-9b89698c6e32\") " Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.380621 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gdv6\" (UniqueName: \"kubernetes.io/projected/571e84f2-a2bc-4f09-ac53-d4a4adafa80b-kube-api-access-2gdv6\") pod \"571e84f2-a2bc-4f09-ac53-d4a4adafa80b\" (UID: \"571e84f2-a2bc-4f09-ac53-d4a4adafa80b\") " Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.380642 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dfbaac71-f99c-4373-a469-f2e5dd0ee632-marketplace-trusted-ca\") pod \"dfbaac71-f99c-4373-a469-f2e5dd0ee632\" (UID: \"dfbaac71-f99c-4373-a469-f2e5dd0ee632\") " Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.381478 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfbaac71-f99c-4373-a469-f2e5dd0ee632-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "dfbaac71-f99c-4373-a469-f2e5dd0ee632" (UID: "dfbaac71-f99c-4373-a469-f2e5dd0ee632"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.382100 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70ad413a-5f81-4094-b2d8-9b89698c6e32-utilities" (OuterVolumeSpecName: "utilities") pod "70ad413a-5f81-4094-b2d8-9b89698c6e32" (UID: "70ad413a-5f81-4094-b2d8-9b89698c6e32"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.382910 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/571e84f2-a2bc-4f09-ac53-d4a4adafa80b-utilities" (OuterVolumeSpecName: "utilities") pod "571e84f2-a2bc-4f09-ac53-d4a4adafa80b" (UID: "571e84f2-a2bc-4f09-ac53-d4a4adafa80b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.385344 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bb65dee-cd5f-46b3-9e7d-36e5d182d19e-utilities" (OuterVolumeSpecName: "utilities") pod "5bb65dee-cd5f-46b3-9e7d-36e5d182d19e" (UID: "5bb65dee-cd5f-46b3-9e7d-36e5d182d19e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.401073 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bb65dee-cd5f-46b3-9e7d-36e5d182d19e-kube-api-access-7ww6h" (OuterVolumeSpecName: "kube-api-access-7ww6h") pod "5bb65dee-cd5f-46b3-9e7d-36e5d182d19e" (UID: "5bb65dee-cd5f-46b3-9e7d-36e5d182d19e"). InnerVolumeSpecName "kube-api-access-7ww6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.404968 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wbtgv"] Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.410303 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wbtgv"] Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.413012 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfbaac71-f99c-4373-a469-f2e5dd0ee632-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "dfbaac71-f99c-4373-a469-f2e5dd0ee632" (UID: "dfbaac71-f99c-4373-a469-f2e5dd0ee632"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.413096 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfbaac71-f99c-4373-a469-f2e5dd0ee632-kube-api-access-m44cw" (OuterVolumeSpecName: "kube-api-access-m44cw") pod "dfbaac71-f99c-4373-a469-f2e5dd0ee632" (UID: "dfbaac71-f99c-4373-a469-f2e5dd0ee632"). InnerVolumeSpecName "kube-api-access-m44cw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.414016 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70ad413a-5f81-4094-b2d8-9b89698c6e32-kube-api-access-vpppf" (OuterVolumeSpecName: "kube-api-access-vpppf") pod "70ad413a-5f81-4094-b2d8-9b89698c6e32" (UID: "70ad413a-5f81-4094-b2d8-9b89698c6e32"). InnerVolumeSpecName "kube-api-access-vpppf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.414237 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/571e84f2-a2bc-4f09-ac53-d4a4adafa80b-kube-api-access-2gdv6" (OuterVolumeSpecName: "kube-api-access-2gdv6") pod "571e84f2-a2bc-4f09-ac53-d4a4adafa80b" (UID: "571e84f2-a2bc-4f09-ac53-d4a4adafa80b"). InnerVolumeSpecName "kube-api-access-2gdv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.414476 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z7fcx"] Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.417539 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/571e84f2-a2bc-4f09-ac53-d4a4adafa80b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "571e84f2-a2bc-4f09-ac53-d4a4adafa80b" (UID: "571e84f2-a2bc-4f09-ac53-d4a4adafa80b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:54:29 crc kubenswrapper[4775]: W0321 04:54:29.418803 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59fec450_4b61_4a15_b1b5_b47dedd649a0.slice/crio-14ec881f38f46e3ad501a36d40d0dbbc592306dd4c1d73a4372f538296655a4f WatchSource:0}: Error finding container 14ec881f38f46e3ad501a36d40d0dbbc592306dd4c1d73a4372f538296655a4f: Status 404 returned error can't find the container with id 14ec881f38f46e3ad501a36d40d0dbbc592306dd4c1d73a4372f538296655a4f Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.422152 4775 scope.go:117] "RemoveContainer" containerID="3b507bd384c25213b4044dd19f72806386309a77448d646faeed8c941611578b" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.439102 4775 scope.go:117] "RemoveContainer" containerID="dd9cc974af699ec57297810c7016806ecad201037353fbcf0dc303241f78fc0b" Mar 21 04:54:29 crc kubenswrapper[4775]: E0321 04:54:29.439598 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd9cc974af699ec57297810c7016806ecad201037353fbcf0dc303241f78fc0b\": container with ID starting with dd9cc974af699ec57297810c7016806ecad201037353fbcf0dc303241f78fc0b not found: ID does not exist" containerID="dd9cc974af699ec57297810c7016806ecad201037353fbcf0dc303241f78fc0b" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.439668 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd9cc974af699ec57297810c7016806ecad201037353fbcf0dc303241f78fc0b"} err="failed to get container status \"dd9cc974af699ec57297810c7016806ecad201037353fbcf0dc303241f78fc0b\": rpc error: code = NotFound desc = could not find container \"dd9cc974af699ec57297810c7016806ecad201037353fbcf0dc303241f78fc0b\": container with ID starting with dd9cc974af699ec57297810c7016806ecad201037353fbcf0dc303241f78fc0b not found: ID does not exist" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.439703 4775 scope.go:117] "RemoveContainer" containerID="1098f33834501f7272dae8532213ab991de93513a5cd5e8028551095da4598b0" Mar 21 04:54:29 crc kubenswrapper[4775]: E0321 04:54:29.440295 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1098f33834501f7272dae8532213ab991de93513a5cd5e8028551095da4598b0\": container with ID starting with 1098f33834501f7272dae8532213ab991de93513a5cd5e8028551095da4598b0 not found: ID does not exist" containerID="1098f33834501f7272dae8532213ab991de93513a5cd5e8028551095da4598b0" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.440358 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1098f33834501f7272dae8532213ab991de93513a5cd5e8028551095da4598b0"} err="failed to get container status \"1098f33834501f7272dae8532213ab991de93513a5cd5e8028551095da4598b0\": rpc error: code = NotFound desc = could not find container \"1098f33834501f7272dae8532213ab991de93513a5cd5e8028551095da4598b0\": container with ID starting with 1098f33834501f7272dae8532213ab991de93513a5cd5e8028551095da4598b0 not found: ID does not exist" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.440391 4775 scope.go:117] "RemoveContainer" containerID="3b507bd384c25213b4044dd19f72806386309a77448d646faeed8c941611578b" Mar 21 04:54:29 crc kubenswrapper[4775]: E0321 04:54:29.440689 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b507bd384c25213b4044dd19f72806386309a77448d646faeed8c941611578b\": container with ID starting with 3b507bd384c25213b4044dd19f72806386309a77448d646faeed8c941611578b not found: ID does not exist" containerID="3b507bd384c25213b4044dd19f72806386309a77448d646faeed8c941611578b" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.440715 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b507bd384c25213b4044dd19f72806386309a77448d646faeed8c941611578b"} err="failed to get container status \"3b507bd384c25213b4044dd19f72806386309a77448d646faeed8c941611578b\": rpc error: code = NotFound desc = could not find container \"3b507bd384c25213b4044dd19f72806386309a77448d646faeed8c941611578b\": container with ID starting with 3b507bd384c25213b4044dd19f72806386309a77448d646faeed8c941611578b not found: ID does not exist" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.440734 4775 scope.go:117] "RemoveContainer" containerID="aaa815a5f0e220d5cd57bb1c4b415350dcf17e31c1b987c38d4e0714c4280f96" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.446602 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bb65dee-cd5f-46b3-9e7d-36e5d182d19e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5bb65dee-cd5f-46b3-9e7d-36e5d182d19e" (UID: "5bb65dee-cd5f-46b3-9e7d-36e5d182d19e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.461400 4775 scope.go:117] "RemoveContainer" containerID="8b5a6baf6223e16aada12bd66d033b18e08ee43831676efdf6aa9934129688fc" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.481934 4775 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dfbaac71-f99c-4373-a469-f2e5dd0ee632-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.481973 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m44cw\" (UniqueName: \"kubernetes.io/projected/dfbaac71-f99c-4373-a469-f2e5dd0ee632-kube-api-access-m44cw\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.481988 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/571e84f2-a2bc-4f09-ac53-d4a4adafa80b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.482057 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bb65dee-cd5f-46b3-9e7d-36e5d182d19e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.482085 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70ad413a-5f81-4094-b2d8-9b89698c6e32-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.482099 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/571e84f2-a2bc-4f09-ac53-d4a4adafa80b-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.482130 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ww6h\" (UniqueName: \"kubernetes.io/projected/5bb65dee-cd5f-46b3-9e7d-36e5d182d19e-kube-api-access-7ww6h\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.482148 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gdv6\" (UniqueName: \"kubernetes.io/projected/571e84f2-a2bc-4f09-ac53-d4a4adafa80b-kube-api-access-2gdv6\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.482164 4775 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dfbaac71-f99c-4373-a469-f2e5dd0ee632-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.482180 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bb65dee-cd5f-46b3-9e7d-36e5d182d19e-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.482195 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpppf\" (UniqueName: \"kubernetes.io/projected/70ad413a-5f81-4094-b2d8-9b89698c6e32-kube-api-access-vpppf\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.489165 4775 scope.go:117] "RemoveContainer" containerID="c8e5e4f89cea3a5393cd7ded536ab2f11744f74388f7645dc484cfc3075f65ca" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.501109 4775 scope.go:117] "RemoveContainer" containerID="aaa815a5f0e220d5cd57bb1c4b415350dcf17e31c1b987c38d4e0714c4280f96" Mar 21 04:54:29 crc kubenswrapper[4775]: E0321 04:54:29.501576 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aaa815a5f0e220d5cd57bb1c4b415350dcf17e31c1b987c38d4e0714c4280f96\": container with ID starting with aaa815a5f0e220d5cd57bb1c4b415350dcf17e31c1b987c38d4e0714c4280f96 not found: ID does not exist" containerID="aaa815a5f0e220d5cd57bb1c4b415350dcf17e31c1b987c38d4e0714c4280f96" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.501617 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaa815a5f0e220d5cd57bb1c4b415350dcf17e31c1b987c38d4e0714c4280f96"} err="failed to get container status \"aaa815a5f0e220d5cd57bb1c4b415350dcf17e31c1b987c38d4e0714c4280f96\": rpc error: code = NotFound desc = could not find container \"aaa815a5f0e220d5cd57bb1c4b415350dcf17e31c1b987c38d4e0714c4280f96\": container with ID starting with aaa815a5f0e220d5cd57bb1c4b415350dcf17e31c1b987c38d4e0714c4280f96 not found: ID does not exist" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.501648 4775 scope.go:117] "RemoveContainer" containerID="8b5a6baf6223e16aada12bd66d033b18e08ee43831676efdf6aa9934129688fc" Mar 21 04:54:29 crc kubenswrapper[4775]: E0321 04:54:29.502046 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b5a6baf6223e16aada12bd66d033b18e08ee43831676efdf6aa9934129688fc\": container with ID starting with 8b5a6baf6223e16aada12bd66d033b18e08ee43831676efdf6aa9934129688fc not found: ID does not exist" containerID="8b5a6baf6223e16aada12bd66d033b18e08ee43831676efdf6aa9934129688fc" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.502079 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b5a6baf6223e16aada12bd66d033b18e08ee43831676efdf6aa9934129688fc"} err="failed to get container status \"8b5a6baf6223e16aada12bd66d033b18e08ee43831676efdf6aa9934129688fc\": rpc error: code = NotFound desc = could not find container \"8b5a6baf6223e16aada12bd66d033b18e08ee43831676efdf6aa9934129688fc\": container with ID starting with 8b5a6baf6223e16aada12bd66d033b18e08ee43831676efdf6aa9934129688fc not found: ID does not exist" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.502102 4775 scope.go:117] "RemoveContainer" containerID="c8e5e4f89cea3a5393cd7ded536ab2f11744f74388f7645dc484cfc3075f65ca" Mar 21 04:54:29 crc kubenswrapper[4775]: E0321 04:54:29.502413 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8e5e4f89cea3a5393cd7ded536ab2f11744f74388f7645dc484cfc3075f65ca\": container with ID starting with c8e5e4f89cea3a5393cd7ded536ab2f11744f74388f7645dc484cfc3075f65ca not found: ID does not exist" containerID="c8e5e4f89cea3a5393cd7ded536ab2f11744f74388f7645dc484cfc3075f65ca" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.502432 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8e5e4f89cea3a5393cd7ded536ab2f11744f74388f7645dc484cfc3075f65ca"} err="failed to get container status \"c8e5e4f89cea3a5393cd7ded536ab2f11744f74388f7645dc484cfc3075f65ca\": rpc error: code = NotFound desc = could not find container \"c8e5e4f89cea3a5393cd7ded536ab2f11744f74388f7645dc484cfc3075f65ca\": container with ID starting with c8e5e4f89cea3a5393cd7ded536ab2f11744f74388f7645dc484cfc3075f65ca not found: ID does not exist" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.502489 4775 scope.go:117] "RemoveContainer" containerID="53e93eb8775f797ec5701d0054608e0cc06d124aa1940ad067f4c33b12a82fa2" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.519889 4775 scope.go:117] "RemoveContainer" containerID="e6613d34aa689f2c33b46ca7fb2fff89fd26737df8f8eb26a52c2030909ae41f" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.537357 4775 scope.go:117] "RemoveContainer" containerID="53e93eb8775f797ec5701d0054608e0cc06d124aa1940ad067f4c33b12a82fa2" Mar 21 04:54:29 crc kubenswrapper[4775]: E0321 04:54:29.538202 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53e93eb8775f797ec5701d0054608e0cc06d124aa1940ad067f4c33b12a82fa2\": container with ID starting with 53e93eb8775f797ec5701d0054608e0cc06d124aa1940ad067f4c33b12a82fa2 not found: ID does not exist" containerID="53e93eb8775f797ec5701d0054608e0cc06d124aa1940ad067f4c33b12a82fa2" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.538313 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53e93eb8775f797ec5701d0054608e0cc06d124aa1940ad067f4c33b12a82fa2"} err="failed to get container status \"53e93eb8775f797ec5701d0054608e0cc06d124aa1940ad067f4c33b12a82fa2\": rpc error: code = NotFound desc = could not find container \"53e93eb8775f797ec5701d0054608e0cc06d124aa1940ad067f4c33b12a82fa2\": container with ID starting with 53e93eb8775f797ec5701d0054608e0cc06d124aa1940ad067f4c33b12a82fa2 not found: ID does not exist" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.538350 4775 scope.go:117] "RemoveContainer" containerID="e6613d34aa689f2c33b46ca7fb2fff89fd26737df8f8eb26a52c2030909ae41f" Mar 21 04:54:29 crc kubenswrapper[4775]: E0321 04:54:29.538629 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6613d34aa689f2c33b46ca7fb2fff89fd26737df8f8eb26a52c2030909ae41f\": container with ID starting with e6613d34aa689f2c33b46ca7fb2fff89fd26737df8f8eb26a52c2030909ae41f not found: ID does not exist" containerID="e6613d34aa689f2c33b46ca7fb2fff89fd26737df8f8eb26a52c2030909ae41f" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.538651 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6613d34aa689f2c33b46ca7fb2fff89fd26737df8f8eb26a52c2030909ae41f"} err="failed to get container status \"e6613d34aa689f2c33b46ca7fb2fff89fd26737df8f8eb26a52c2030909ae41f\": rpc error: code = NotFound desc = could not find container \"e6613d34aa689f2c33b46ca7fb2fff89fd26737df8f8eb26a52c2030909ae41f\": container with ID starting with e6613d34aa689f2c33b46ca7fb2fff89fd26737df8f8eb26a52c2030909ae41f not found: ID does not exist" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.538665 4775 scope.go:117] "RemoveContainer" containerID="bce5cf8972de39d9de712cead6829664ca56252c15f4837eb08ebb6d9c3a039b" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.545846 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70ad413a-5f81-4094-b2d8-9b89698c6e32-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "70ad413a-5f81-4094-b2d8-9b89698c6e32" (UID: "70ad413a-5f81-4094-b2d8-9b89698c6e32"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.553567 4775 scope.go:117] "RemoveContainer" containerID="a669beb6a121e654078eda21a56b0d2171a5e04b1968ec341bfe87375574ff9a" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.583822 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70ad413a-5f81-4094-b2d8-9b89698c6e32-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.607435 4775 scope.go:117] "RemoveContainer" containerID="ee5a58cbfa0edd60aaa49923480f7f0b63c1139c1f29a46a06572d11e904c5be" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.637418 4775 scope.go:117] "RemoveContainer" containerID="bce5cf8972de39d9de712cead6829664ca56252c15f4837eb08ebb6d9c3a039b" Mar 21 04:54:29 crc kubenswrapper[4775]: E0321 04:54:29.637960 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bce5cf8972de39d9de712cead6829664ca56252c15f4837eb08ebb6d9c3a039b\": container with ID starting with bce5cf8972de39d9de712cead6829664ca56252c15f4837eb08ebb6d9c3a039b not found: ID does not exist" containerID="bce5cf8972de39d9de712cead6829664ca56252c15f4837eb08ebb6d9c3a039b" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.638012 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bce5cf8972de39d9de712cead6829664ca56252c15f4837eb08ebb6d9c3a039b"} err="failed to get container status \"bce5cf8972de39d9de712cead6829664ca56252c15f4837eb08ebb6d9c3a039b\": rpc error: code = NotFound desc = could not find container \"bce5cf8972de39d9de712cead6829664ca56252c15f4837eb08ebb6d9c3a039b\": container with ID starting with bce5cf8972de39d9de712cead6829664ca56252c15f4837eb08ebb6d9c3a039b not found: ID does not exist" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.638047 4775 scope.go:117] "RemoveContainer" containerID="a669beb6a121e654078eda21a56b0d2171a5e04b1968ec341bfe87375574ff9a" Mar 21 04:54:29 crc kubenswrapper[4775]: E0321 04:54:29.638397 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a669beb6a121e654078eda21a56b0d2171a5e04b1968ec341bfe87375574ff9a\": container with ID starting with a669beb6a121e654078eda21a56b0d2171a5e04b1968ec341bfe87375574ff9a not found: ID does not exist" containerID="a669beb6a121e654078eda21a56b0d2171a5e04b1968ec341bfe87375574ff9a" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.638424 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a669beb6a121e654078eda21a56b0d2171a5e04b1968ec341bfe87375574ff9a"} err="failed to get container status \"a669beb6a121e654078eda21a56b0d2171a5e04b1968ec341bfe87375574ff9a\": rpc error: code = NotFound desc = could not find container \"a669beb6a121e654078eda21a56b0d2171a5e04b1968ec341bfe87375574ff9a\": container with ID starting with a669beb6a121e654078eda21a56b0d2171a5e04b1968ec341bfe87375574ff9a not found: ID does not exist" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.638440 4775 scope.go:117] "RemoveContainer" containerID="ee5a58cbfa0edd60aaa49923480f7f0b63c1139c1f29a46a06572d11e904c5be" Mar 21 04:54:29 crc kubenswrapper[4775]: E0321 04:54:29.638762 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee5a58cbfa0edd60aaa49923480f7f0b63c1139c1f29a46a06572d11e904c5be\": container with ID starting with ee5a58cbfa0edd60aaa49923480f7f0b63c1139c1f29a46a06572d11e904c5be not found: ID does not exist" containerID="ee5a58cbfa0edd60aaa49923480f7f0b63c1139c1f29a46a06572d11e904c5be" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.638799 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee5a58cbfa0edd60aaa49923480f7f0b63c1139c1f29a46a06572d11e904c5be"} err="failed to get container status \"ee5a58cbfa0edd60aaa49923480f7f0b63c1139c1f29a46a06572d11e904c5be\": rpc error: code = NotFound desc = could not find container \"ee5a58cbfa0edd60aaa49923480f7f0b63c1139c1f29a46a06572d11e904c5be\": container with ID starting with ee5a58cbfa0edd60aaa49923480f7f0b63c1139c1f29a46a06572d11e904c5be not found: ID does not exist" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.638815 4775 scope.go:117] "RemoveContainer" containerID="02ab090064cf43714a361f98c8401a08ff5684170017c160d3affd080275fc5a" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.657093 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wjfwd"] Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.661482 4775 scope.go:117] "RemoveContainer" containerID="21eb68b1dc783b9a4153f39746a113d0c901b8fcfacb75d4fa0041622f71eb2d" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.674509 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" path="/var/lib/kubelet/pods/ccb910ab-dcef-4523-81df-c0fb5eb83429/volumes" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.677702 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wjfwd"] Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.683636 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-959kj"] Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.688509 4775 scope.go:117] "RemoveContainer" containerID="9f77f33cd00f43111cd055698fe7da46bda8e9c64a949491d07dadf580cf4426" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.696975 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-959kj"] Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.701537 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wrzs4"] Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.705102 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wrzs4"] Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.708837 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hldm7"] Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.709082 4775 scope.go:117] "RemoveContainer" containerID="02ab090064cf43714a361f98c8401a08ff5684170017c160d3affd080275fc5a" Mar 21 04:54:29 crc kubenswrapper[4775]: E0321 04:54:29.709548 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02ab090064cf43714a361f98c8401a08ff5684170017c160d3affd080275fc5a\": container with ID starting with 02ab090064cf43714a361f98c8401a08ff5684170017c160d3affd080275fc5a not found: ID does not exist" containerID="02ab090064cf43714a361f98c8401a08ff5684170017c160d3affd080275fc5a" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.709608 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02ab090064cf43714a361f98c8401a08ff5684170017c160d3affd080275fc5a"} err="failed to get container status \"02ab090064cf43714a361f98c8401a08ff5684170017c160d3affd080275fc5a\": rpc error: code = NotFound desc = could not find container \"02ab090064cf43714a361f98c8401a08ff5684170017c160d3affd080275fc5a\": container with ID starting with 02ab090064cf43714a361f98c8401a08ff5684170017c160d3affd080275fc5a not found: ID does not exist" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.709644 4775 scope.go:117] "RemoveContainer" containerID="21eb68b1dc783b9a4153f39746a113d0c901b8fcfacb75d4fa0041622f71eb2d" Mar 21 04:54:29 crc kubenswrapper[4775]: E0321 04:54:29.710061 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21eb68b1dc783b9a4153f39746a113d0c901b8fcfacb75d4fa0041622f71eb2d\": container with ID starting with 21eb68b1dc783b9a4153f39746a113d0c901b8fcfacb75d4fa0041622f71eb2d not found: ID does not exist" containerID="21eb68b1dc783b9a4153f39746a113d0c901b8fcfacb75d4fa0041622f71eb2d" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.710093 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21eb68b1dc783b9a4153f39746a113d0c901b8fcfacb75d4fa0041622f71eb2d"} err="failed to get container status \"21eb68b1dc783b9a4153f39746a113d0c901b8fcfacb75d4fa0041622f71eb2d\": rpc error: code = NotFound desc = could not find container \"21eb68b1dc783b9a4153f39746a113d0c901b8fcfacb75d4fa0041622f71eb2d\": container with ID starting with 21eb68b1dc783b9a4153f39746a113d0c901b8fcfacb75d4fa0041622f71eb2d not found: ID does not exist" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.710125 4775 scope.go:117] "RemoveContainer" containerID="9f77f33cd00f43111cd055698fe7da46bda8e9c64a949491d07dadf580cf4426" Mar 21 04:54:29 crc kubenswrapper[4775]: E0321 04:54:29.710601 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f77f33cd00f43111cd055698fe7da46bda8e9c64a949491d07dadf580cf4426\": container with ID starting with 9f77f33cd00f43111cd055698fe7da46bda8e9c64a949491d07dadf580cf4426 not found: ID does not exist" containerID="9f77f33cd00f43111cd055698fe7da46bda8e9c64a949491d07dadf580cf4426" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.710623 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f77f33cd00f43111cd055698fe7da46bda8e9c64a949491d07dadf580cf4426"} err="failed to get container status \"9f77f33cd00f43111cd055698fe7da46bda8e9c64a949491d07dadf580cf4426\": rpc error: code = NotFound desc = could not find container \"9f77f33cd00f43111cd055698fe7da46bda8e9c64a949491d07dadf580cf4426\": container with ID starting with 9f77f33cd00f43111cd055698fe7da46bda8e9c64a949491d07dadf580cf4426 not found: ID does not exist" Mar 21 04:54:29 crc kubenswrapper[4775]: I0321 04:54:29.711780 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hldm7"] Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.335511 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z7fcx" event={"ID":"59fec450-4b61-4a15-b1b5-b47dedd649a0","Type":"ContainerStarted","Data":"cca7682b1f9e0c6812bda17dc39c9c1c9aaa1c03a18dac4025f41b93dff335ff"} Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.335549 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z7fcx" event={"ID":"59fec450-4b61-4a15-b1b5-b47dedd649a0","Type":"ContainerStarted","Data":"14ec881f38f46e3ad501a36d40d0dbbc592306dd4c1d73a4372f538296655a4f"} Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.335779 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-z7fcx" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.338867 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-z7fcx" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.367875 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-z7fcx" podStartSLOduration=2.3678618289999998 podStartE2EDuration="2.367861829s" podCreationTimestamp="2026-03-21 04:54:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 04:54:30.351801832 +0000 UTC m=+423.328265486" watchObservedRunningTime="2026-03-21 04:54:30.367861829 +0000 UTC m=+423.344325453" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.439376 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-77cp9"] Mar 21 04:54:30 crc kubenswrapper[4775]: E0321 04:54:30.439605 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bb65dee-cd5f-46b3-9e7d-36e5d182d19e" containerName="extract-content" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.439620 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bb65dee-cd5f-46b3-9e7d-36e5d182d19e" containerName="extract-content" Mar 21 04:54:30 crc kubenswrapper[4775]: E0321 04:54:30.439635 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70ad413a-5f81-4094-b2d8-9b89698c6e32" containerName="registry-server" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.439643 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="70ad413a-5f81-4094-b2d8-9b89698c6e32" containerName="registry-server" Mar 21 04:54:30 crc kubenswrapper[4775]: E0321 04:54:30.439656 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" containerName="extract-utilities" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.439665 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" containerName="extract-utilities" Mar 21 04:54:30 crc kubenswrapper[4775]: E0321 04:54:30.439679 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" containerName="extract-content" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.439686 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" containerName="extract-content" Mar 21 04:54:30 crc kubenswrapper[4775]: E0321 04:54:30.439698 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70ad413a-5f81-4094-b2d8-9b89698c6e32" containerName="extract-content" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.439705 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="70ad413a-5f81-4094-b2d8-9b89698c6e32" containerName="extract-content" Mar 21 04:54:30 crc kubenswrapper[4775]: E0321 04:54:30.439716 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfbaac71-f99c-4373-a469-f2e5dd0ee632" containerName="marketplace-operator" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.439724 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfbaac71-f99c-4373-a469-f2e5dd0ee632" containerName="marketplace-operator" Mar 21 04:54:30 crc kubenswrapper[4775]: E0321 04:54:30.439735 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" containerName="registry-server" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.439742 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" containerName="registry-server" Mar 21 04:54:30 crc kubenswrapper[4775]: E0321 04:54:30.439751 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" containerName="extract-utilities" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.439758 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" containerName="extract-utilities" Mar 21 04:54:30 crc kubenswrapper[4775]: E0321 04:54:30.439771 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfbaac71-f99c-4373-a469-f2e5dd0ee632" containerName="marketplace-operator" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.439778 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfbaac71-f99c-4373-a469-f2e5dd0ee632" containerName="marketplace-operator" Mar 21 04:54:30 crc kubenswrapper[4775]: E0321 04:54:30.439789 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bb65dee-cd5f-46b3-9e7d-36e5d182d19e" containerName="extract-utilities" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.439795 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bb65dee-cd5f-46b3-9e7d-36e5d182d19e" containerName="extract-utilities" Mar 21 04:54:30 crc kubenswrapper[4775]: E0321 04:54:30.439805 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" containerName="registry-server" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.439812 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" containerName="registry-server" Mar 21 04:54:30 crc kubenswrapper[4775]: E0321 04:54:30.439822 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70ad413a-5f81-4094-b2d8-9b89698c6e32" containerName="extract-utilities" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.439829 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="70ad413a-5f81-4094-b2d8-9b89698c6e32" containerName="extract-utilities" Mar 21 04:54:30 crc kubenswrapper[4775]: E0321 04:54:30.439838 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bb65dee-cd5f-46b3-9e7d-36e5d182d19e" containerName="registry-server" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.439844 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bb65dee-cd5f-46b3-9e7d-36e5d182d19e" containerName="registry-server" Mar 21 04:54:30 crc kubenswrapper[4775]: E0321 04:54:30.439856 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" containerName="extract-content" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.439863 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" containerName="extract-content" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.439963 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccb910ab-dcef-4523-81df-c0fb5eb83429" containerName="registry-server" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.439974 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="70ad413a-5f81-4094-b2d8-9b89698c6e32" containerName="registry-server" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.439985 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bb65dee-cd5f-46b3-9e7d-36e5d182d19e" containerName="registry-server" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.439994 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" containerName="registry-server" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.440002 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfbaac71-f99c-4373-a469-f2e5dd0ee632" containerName="marketplace-operator" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.440015 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfbaac71-f99c-4373-a469-f2e5dd0ee632" containerName="marketplace-operator" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.440878 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-77cp9" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.446083 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.483367 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-77cp9"] Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.495489 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6dhm\" (UniqueName: \"kubernetes.io/projected/08558c65-4599-4c64-bb0f-f18f94cecdec-kube-api-access-q6dhm\") pod \"certified-operators-77cp9\" (UID: \"08558c65-4599-4c64-bb0f-f18f94cecdec\") " pod="openshift-marketplace/certified-operators-77cp9" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.495559 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08558c65-4599-4c64-bb0f-f18f94cecdec-catalog-content\") pod \"certified-operators-77cp9\" (UID: \"08558c65-4599-4c64-bb0f-f18f94cecdec\") " pod="openshift-marketplace/certified-operators-77cp9" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.495733 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08558c65-4599-4c64-bb0f-f18f94cecdec-utilities\") pod \"certified-operators-77cp9\" (UID: \"08558c65-4599-4c64-bb0f-f18f94cecdec\") " pod="openshift-marketplace/certified-operators-77cp9" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.596993 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08558c65-4599-4c64-bb0f-f18f94cecdec-utilities\") pod \"certified-operators-77cp9\" (UID: \"08558c65-4599-4c64-bb0f-f18f94cecdec\") " pod="openshift-marketplace/certified-operators-77cp9" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.597078 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6dhm\" (UniqueName: \"kubernetes.io/projected/08558c65-4599-4c64-bb0f-f18f94cecdec-kube-api-access-q6dhm\") pod \"certified-operators-77cp9\" (UID: \"08558c65-4599-4c64-bb0f-f18f94cecdec\") " pod="openshift-marketplace/certified-operators-77cp9" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.597147 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08558c65-4599-4c64-bb0f-f18f94cecdec-catalog-content\") pod \"certified-operators-77cp9\" (UID: \"08558c65-4599-4c64-bb0f-f18f94cecdec\") " pod="openshift-marketplace/certified-operators-77cp9" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.597977 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08558c65-4599-4c64-bb0f-f18f94cecdec-utilities\") pod \"certified-operators-77cp9\" (UID: \"08558c65-4599-4c64-bb0f-f18f94cecdec\") " pod="openshift-marketplace/certified-operators-77cp9" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.598034 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08558c65-4599-4c64-bb0f-f18f94cecdec-catalog-content\") pod \"certified-operators-77cp9\" (UID: \"08558c65-4599-4c64-bb0f-f18f94cecdec\") " pod="openshift-marketplace/certified-operators-77cp9" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.616054 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6dhm\" (UniqueName: \"kubernetes.io/projected/08558c65-4599-4c64-bb0f-f18f94cecdec-kube-api-access-q6dhm\") pod \"certified-operators-77cp9\" (UID: \"08558c65-4599-4c64-bb0f-f18f94cecdec\") " pod="openshift-marketplace/certified-operators-77cp9" Mar 21 04:54:30 crc kubenswrapper[4775]: I0321 04:54:30.797354 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-77cp9" Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.042775 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h4wmk"] Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.043702 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4wmk" Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.045654 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.053757 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4wmk"] Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.102269 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a9e7d52-67b3-4a38-a978-6566b1c7870a-catalog-content\") pod \"redhat-marketplace-h4wmk\" (UID: \"4a9e7d52-67b3-4a38-a978-6566b1c7870a\") " pod="openshift-marketplace/redhat-marketplace-h4wmk" Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.102339 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a9e7d52-67b3-4a38-a978-6566b1c7870a-utilities\") pod \"redhat-marketplace-h4wmk\" (UID: \"4a9e7d52-67b3-4a38-a978-6566b1c7870a\") " pod="openshift-marketplace/redhat-marketplace-h4wmk" Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.102404 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzhv8\" (UniqueName: \"kubernetes.io/projected/4a9e7d52-67b3-4a38-a978-6566b1c7870a-kube-api-access-hzhv8\") pod \"redhat-marketplace-h4wmk\" (UID: \"4a9e7d52-67b3-4a38-a978-6566b1c7870a\") " pod="openshift-marketplace/redhat-marketplace-h4wmk" Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.183688 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-77cp9"] Mar 21 04:54:31 crc kubenswrapper[4775]: W0321 04:54:31.189794 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08558c65_4599_4c64_bb0f_f18f94cecdec.slice/crio-53d8511c66dacf6ae929a89765dca05bde8c11649a74ae9891aa53b8786e1f13 WatchSource:0}: Error finding container 53d8511c66dacf6ae929a89765dca05bde8c11649a74ae9891aa53b8786e1f13: Status 404 returned error can't find the container with id 53d8511c66dacf6ae929a89765dca05bde8c11649a74ae9891aa53b8786e1f13 Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.203061 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzhv8\" (UniqueName: \"kubernetes.io/projected/4a9e7d52-67b3-4a38-a978-6566b1c7870a-kube-api-access-hzhv8\") pod \"redhat-marketplace-h4wmk\" (UID: \"4a9e7d52-67b3-4a38-a978-6566b1c7870a\") " pod="openshift-marketplace/redhat-marketplace-h4wmk" Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.203440 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a9e7d52-67b3-4a38-a978-6566b1c7870a-catalog-content\") pod \"redhat-marketplace-h4wmk\" (UID: \"4a9e7d52-67b3-4a38-a978-6566b1c7870a\") " pod="openshift-marketplace/redhat-marketplace-h4wmk" Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.203481 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a9e7d52-67b3-4a38-a978-6566b1c7870a-utilities\") pod \"redhat-marketplace-h4wmk\" (UID: \"4a9e7d52-67b3-4a38-a978-6566b1c7870a\") " pod="openshift-marketplace/redhat-marketplace-h4wmk" Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.203941 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a9e7d52-67b3-4a38-a978-6566b1c7870a-utilities\") pod \"redhat-marketplace-h4wmk\" (UID: \"4a9e7d52-67b3-4a38-a978-6566b1c7870a\") " pod="openshift-marketplace/redhat-marketplace-h4wmk" Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.203978 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a9e7d52-67b3-4a38-a978-6566b1c7870a-catalog-content\") pod \"redhat-marketplace-h4wmk\" (UID: \"4a9e7d52-67b3-4a38-a978-6566b1c7870a\") " pod="openshift-marketplace/redhat-marketplace-h4wmk" Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.222107 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzhv8\" (UniqueName: \"kubernetes.io/projected/4a9e7d52-67b3-4a38-a978-6566b1c7870a-kube-api-access-hzhv8\") pod \"redhat-marketplace-h4wmk\" (UID: \"4a9e7d52-67b3-4a38-a978-6566b1c7870a\") " pod="openshift-marketplace/redhat-marketplace-h4wmk" Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.348160 4775 generic.go:334] "Generic (PLEG): container finished" podID="08558c65-4599-4c64-bb0f-f18f94cecdec" containerID="c1c5aff8dc874c9ce56fbe3e84035b64de894d7afcf43417e89df45fea4306ba" exitCode=0 Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.348255 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-77cp9" event={"ID":"08558c65-4599-4c64-bb0f-f18f94cecdec","Type":"ContainerDied","Data":"c1c5aff8dc874c9ce56fbe3e84035b64de894d7afcf43417e89df45fea4306ba"} Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.348310 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-77cp9" event={"ID":"08558c65-4599-4c64-bb0f-f18f94cecdec","Type":"ContainerStarted","Data":"53d8511c66dacf6ae929a89765dca05bde8c11649a74ae9891aa53b8786e1f13"} Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.370094 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4wmk" Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.672295 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="571e84f2-a2bc-4f09-ac53-d4a4adafa80b" path="/var/lib/kubelet/pods/571e84f2-a2bc-4f09-ac53-d4a4adafa80b/volumes" Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.673195 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bb65dee-cd5f-46b3-9e7d-36e5d182d19e" path="/var/lib/kubelet/pods/5bb65dee-cd5f-46b3-9e7d-36e5d182d19e/volumes" Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.673963 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70ad413a-5f81-4094-b2d8-9b89698c6e32" path="/var/lib/kubelet/pods/70ad413a-5f81-4094-b2d8-9b89698c6e32/volumes" Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.675043 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfbaac71-f99c-4373-a469-f2e5dd0ee632" path="/var/lib/kubelet/pods/dfbaac71-f99c-4373-a469-f2e5dd0ee632/volumes" Mar 21 04:54:31 crc kubenswrapper[4775]: I0321 04:54:31.746160 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4wmk"] Mar 21 04:54:31 crc kubenswrapper[4775]: W0321 04:54:31.752835 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a9e7d52_67b3_4a38_a978_6566b1c7870a.slice/crio-89817b57585c98f6ed285ff85f3fa80e94e56cc7b45cf060291a6b8ce625b50b WatchSource:0}: Error finding container 89817b57585c98f6ed285ff85f3fa80e94e56cc7b45cf060291a6b8ce625b50b: Status 404 returned error can't find the container with id 89817b57585c98f6ed285ff85f3fa80e94e56cc7b45cf060291a6b8ce625b50b Mar 21 04:54:32 crc kubenswrapper[4775]: I0321 04:54:32.354625 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-77cp9" event={"ID":"08558c65-4599-4c64-bb0f-f18f94cecdec","Type":"ContainerStarted","Data":"8adc4d44e75efefa651623f47e19f981358f6762d49aaf60f4f2e126668c06b9"} Mar 21 04:54:32 crc kubenswrapper[4775]: I0321 04:54:32.355780 4775 generic.go:334] "Generic (PLEG): container finished" podID="4a9e7d52-67b3-4a38-a978-6566b1c7870a" containerID="713719d27d87bbfcfa4cd0e037e75a5bcdb5f83d1603e453d967d0244b5bf4de" exitCode=0 Mar 21 04:54:32 crc kubenswrapper[4775]: I0321 04:54:32.355996 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4wmk" event={"ID":"4a9e7d52-67b3-4a38-a978-6566b1c7870a","Type":"ContainerDied","Data":"713719d27d87bbfcfa4cd0e037e75a5bcdb5f83d1603e453d967d0244b5bf4de"} Mar 21 04:54:32 crc kubenswrapper[4775]: I0321 04:54:32.356037 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4wmk" event={"ID":"4a9e7d52-67b3-4a38-a978-6566b1c7870a","Type":"ContainerStarted","Data":"89817b57585c98f6ed285ff85f3fa80e94e56cc7b45cf060291a6b8ce625b50b"} Mar 21 04:54:32 crc kubenswrapper[4775]: I0321 04:54:32.481878 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 04:54:32 crc kubenswrapper[4775]: I0321 04:54:32.481942 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 04:54:32 crc kubenswrapper[4775]: I0321 04:54:32.848234 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gslhk"] Mar 21 04:54:32 crc kubenswrapper[4775]: I0321 04:54:32.849736 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gslhk" Mar 21 04:54:32 crc kubenswrapper[4775]: I0321 04:54:32.851933 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 21 04:54:32 crc kubenswrapper[4775]: I0321 04:54:32.854346 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gslhk"] Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.024228 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9p7c\" (UniqueName: \"kubernetes.io/projected/373fa4a4-80b8-4b32-a9ee-b272604d3adc-kube-api-access-h9p7c\") pod \"redhat-operators-gslhk\" (UID: \"373fa4a4-80b8-4b32-a9ee-b272604d3adc\") " pod="openshift-marketplace/redhat-operators-gslhk" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.024282 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/373fa4a4-80b8-4b32-a9ee-b272604d3adc-utilities\") pod \"redhat-operators-gslhk\" (UID: \"373fa4a4-80b8-4b32-a9ee-b272604d3adc\") " pod="openshift-marketplace/redhat-operators-gslhk" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.024445 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/373fa4a4-80b8-4b32-a9ee-b272604d3adc-catalog-content\") pod \"redhat-operators-gslhk\" (UID: \"373fa4a4-80b8-4b32-a9ee-b272604d3adc\") " pod="openshift-marketplace/redhat-operators-gslhk" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.125953 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9p7c\" (UniqueName: \"kubernetes.io/projected/373fa4a4-80b8-4b32-a9ee-b272604d3adc-kube-api-access-h9p7c\") pod \"redhat-operators-gslhk\" (UID: \"373fa4a4-80b8-4b32-a9ee-b272604d3adc\") " pod="openshift-marketplace/redhat-operators-gslhk" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.126010 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/373fa4a4-80b8-4b32-a9ee-b272604d3adc-utilities\") pod \"redhat-operators-gslhk\" (UID: \"373fa4a4-80b8-4b32-a9ee-b272604d3adc\") " pod="openshift-marketplace/redhat-operators-gslhk" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.126083 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/373fa4a4-80b8-4b32-a9ee-b272604d3adc-catalog-content\") pod \"redhat-operators-gslhk\" (UID: \"373fa4a4-80b8-4b32-a9ee-b272604d3adc\") " pod="openshift-marketplace/redhat-operators-gslhk" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.126547 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/373fa4a4-80b8-4b32-a9ee-b272604d3adc-catalog-content\") pod \"redhat-operators-gslhk\" (UID: \"373fa4a4-80b8-4b32-a9ee-b272604d3adc\") " pod="openshift-marketplace/redhat-operators-gslhk" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.126725 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/373fa4a4-80b8-4b32-a9ee-b272604d3adc-utilities\") pod \"redhat-operators-gslhk\" (UID: \"373fa4a4-80b8-4b32-a9ee-b272604d3adc\") " pod="openshift-marketplace/redhat-operators-gslhk" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.159901 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9p7c\" (UniqueName: \"kubernetes.io/projected/373fa4a4-80b8-4b32-a9ee-b272604d3adc-kube-api-access-h9p7c\") pod \"redhat-operators-gslhk\" (UID: \"373fa4a4-80b8-4b32-a9ee-b272604d3adc\") " pod="openshift-marketplace/redhat-operators-gslhk" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.175180 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gslhk" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.384109 4775 generic.go:334] "Generic (PLEG): container finished" podID="4a9e7d52-67b3-4a38-a978-6566b1c7870a" containerID="64eb994ce0af079d74aaa5eb9da00137399bcf89c4c62a054ea446143999c5d6" exitCode=0 Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.384460 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4wmk" event={"ID":"4a9e7d52-67b3-4a38-a978-6566b1c7870a","Type":"ContainerDied","Data":"64eb994ce0af079d74aaa5eb9da00137399bcf89c4c62a054ea446143999c5d6"} Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.388235 4775 generic.go:334] "Generic (PLEG): container finished" podID="08558c65-4599-4c64-bb0f-f18f94cecdec" containerID="8adc4d44e75efefa651623f47e19f981358f6762d49aaf60f4f2e126668c06b9" exitCode=0 Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.388266 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-77cp9" event={"ID":"08558c65-4599-4c64-bb0f-f18f94cecdec","Type":"ContainerDied","Data":"8adc4d44e75efefa651623f47e19f981358f6762d49aaf60f4f2e126668c06b9"} Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.438830 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x59zn"] Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.439919 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x59zn" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.442149 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.454649 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x59zn"] Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.549005 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gslhk"] Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.631718 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz69p\" (UniqueName: \"kubernetes.io/projected/e8513867-3729-4d93-b8ca-45ecb69c50e6-kube-api-access-xz69p\") pod \"community-operators-x59zn\" (UID: \"e8513867-3729-4d93-b8ca-45ecb69c50e6\") " pod="openshift-marketplace/community-operators-x59zn" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.631769 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8513867-3729-4d93-b8ca-45ecb69c50e6-utilities\") pod \"community-operators-x59zn\" (UID: \"e8513867-3729-4d93-b8ca-45ecb69c50e6\") " pod="openshift-marketplace/community-operators-x59zn" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.631806 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8513867-3729-4d93-b8ca-45ecb69c50e6-catalog-content\") pod \"community-operators-x59zn\" (UID: \"e8513867-3729-4d93-b8ca-45ecb69c50e6\") " pod="openshift-marketplace/community-operators-x59zn" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.733361 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz69p\" (UniqueName: \"kubernetes.io/projected/e8513867-3729-4d93-b8ca-45ecb69c50e6-kube-api-access-xz69p\") pod \"community-operators-x59zn\" (UID: \"e8513867-3729-4d93-b8ca-45ecb69c50e6\") " pod="openshift-marketplace/community-operators-x59zn" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.733444 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8513867-3729-4d93-b8ca-45ecb69c50e6-utilities\") pod \"community-operators-x59zn\" (UID: \"e8513867-3729-4d93-b8ca-45ecb69c50e6\") " pod="openshift-marketplace/community-operators-x59zn" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.733543 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8513867-3729-4d93-b8ca-45ecb69c50e6-catalog-content\") pod \"community-operators-x59zn\" (UID: \"e8513867-3729-4d93-b8ca-45ecb69c50e6\") " pod="openshift-marketplace/community-operators-x59zn" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.733960 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8513867-3729-4d93-b8ca-45ecb69c50e6-utilities\") pod \"community-operators-x59zn\" (UID: \"e8513867-3729-4d93-b8ca-45ecb69c50e6\") " pod="openshift-marketplace/community-operators-x59zn" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.734488 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8513867-3729-4d93-b8ca-45ecb69c50e6-catalog-content\") pod \"community-operators-x59zn\" (UID: \"e8513867-3729-4d93-b8ca-45ecb69c50e6\") " pod="openshift-marketplace/community-operators-x59zn" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.752664 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz69p\" (UniqueName: \"kubernetes.io/projected/e8513867-3729-4d93-b8ca-45ecb69c50e6-kube-api-access-xz69p\") pod \"community-operators-x59zn\" (UID: \"e8513867-3729-4d93-b8ca-45ecb69c50e6\") " pod="openshift-marketplace/community-operators-x59zn" Mar 21 04:54:33 crc kubenswrapper[4775]: I0321 04:54:33.758215 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x59zn" Mar 21 04:54:34 crc kubenswrapper[4775]: I0321 04:54:34.185228 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x59zn"] Mar 21 04:54:34 crc kubenswrapper[4775]: W0321 04:54:34.199276 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8513867_3729_4d93_b8ca_45ecb69c50e6.slice/crio-97fef89995b7abeabce972f4dd29e44dba02fd1652df55c292b81bae7cf2ee8a WatchSource:0}: Error finding container 97fef89995b7abeabce972f4dd29e44dba02fd1652df55c292b81bae7cf2ee8a: Status 404 returned error can't find the container with id 97fef89995b7abeabce972f4dd29e44dba02fd1652df55c292b81bae7cf2ee8a Mar 21 04:54:34 crc kubenswrapper[4775]: I0321 04:54:34.404334 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4wmk" event={"ID":"4a9e7d52-67b3-4a38-a978-6566b1c7870a","Type":"ContainerStarted","Data":"cc1ad29ec9a597dea032729d906af4bdb7605800ee74cf73471404f8b9a6d900"} Mar 21 04:54:34 crc kubenswrapper[4775]: I0321 04:54:34.407457 4775 generic.go:334] "Generic (PLEG): container finished" podID="373fa4a4-80b8-4b32-a9ee-b272604d3adc" containerID="c99df84a5860f15fae9b8f3adb088d68432645354dfffca2f6de63a527e05f0e" exitCode=0 Mar 21 04:54:34 crc kubenswrapper[4775]: I0321 04:54:34.407543 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gslhk" event={"ID":"373fa4a4-80b8-4b32-a9ee-b272604d3adc","Type":"ContainerDied","Data":"c99df84a5860f15fae9b8f3adb088d68432645354dfffca2f6de63a527e05f0e"} Mar 21 04:54:34 crc kubenswrapper[4775]: I0321 04:54:34.407571 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gslhk" event={"ID":"373fa4a4-80b8-4b32-a9ee-b272604d3adc","Type":"ContainerStarted","Data":"e54391c9eb367f896efa6df1cd7ef84b9610e98a86ff539f47326508cb83fcf3"} Mar 21 04:54:34 crc kubenswrapper[4775]: I0321 04:54:34.410181 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-77cp9" event={"ID":"08558c65-4599-4c64-bb0f-f18f94cecdec","Type":"ContainerStarted","Data":"868c4b07f4de3727b8ae04aba6da0fe9b79906be41d2cd1280a77062daa256dd"} Mar 21 04:54:34 crc kubenswrapper[4775]: I0321 04:54:34.414424 4775 generic.go:334] "Generic (PLEG): container finished" podID="e8513867-3729-4d93-b8ca-45ecb69c50e6" containerID="5d48c18f3faab589bc7d3fc497a7cf0764dd1a7869444eb3720b010a70b870c2" exitCode=0 Mar 21 04:54:34 crc kubenswrapper[4775]: I0321 04:54:34.414480 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x59zn" event={"ID":"e8513867-3729-4d93-b8ca-45ecb69c50e6","Type":"ContainerDied","Data":"5d48c18f3faab589bc7d3fc497a7cf0764dd1a7869444eb3720b010a70b870c2"} Mar 21 04:54:34 crc kubenswrapper[4775]: I0321 04:54:34.414539 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x59zn" event={"ID":"e8513867-3729-4d93-b8ca-45ecb69c50e6","Type":"ContainerStarted","Data":"97fef89995b7abeabce972f4dd29e44dba02fd1652df55c292b81bae7cf2ee8a"} Mar 21 04:54:34 crc kubenswrapper[4775]: I0321 04:54:34.430908 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h4wmk" podStartSLOduration=1.926252817 podStartE2EDuration="3.430891419s" podCreationTimestamp="2026-03-21 04:54:31 +0000 UTC" firstStartedPulling="2026-03-21 04:54:32.362665891 +0000 UTC m=+425.339129515" lastFinishedPulling="2026-03-21 04:54:33.867304493 +0000 UTC m=+426.843768117" observedRunningTime="2026-03-21 04:54:34.429537203 +0000 UTC m=+427.406000827" watchObservedRunningTime="2026-03-21 04:54:34.430891419 +0000 UTC m=+427.407355043" Mar 21 04:54:34 crc kubenswrapper[4775]: I0321 04:54:34.508457 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-77cp9" podStartSLOduration=2.060206366 podStartE2EDuration="4.508436942s" podCreationTimestamp="2026-03-21 04:54:30 +0000 UTC" firstStartedPulling="2026-03-21 04:54:31.349333142 +0000 UTC m=+424.325796766" lastFinishedPulling="2026-03-21 04:54:33.797563718 +0000 UTC m=+426.774027342" observedRunningTime="2026-03-21 04:54:34.507673441 +0000 UTC m=+427.484137065" watchObservedRunningTime="2026-03-21 04:54:34.508436942 +0000 UTC m=+427.484900566" Mar 21 04:54:35 crc kubenswrapper[4775]: I0321 04:54:35.421173 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gslhk" event={"ID":"373fa4a4-80b8-4b32-a9ee-b272604d3adc","Type":"ContainerStarted","Data":"872ab07c719410279fd52976d7026dd1e51fa3df57a56a17fe5312e830ff9260"} Mar 21 04:54:35 crc kubenswrapper[4775]: I0321 04:54:35.422859 4775 generic.go:334] "Generic (PLEG): container finished" podID="e8513867-3729-4d93-b8ca-45ecb69c50e6" containerID="ec1634b3909c3b128dbf98f5658155ac371c891b6e1b1b64672b4b0ba678b266" exitCode=0 Mar 21 04:54:35 crc kubenswrapper[4775]: I0321 04:54:35.423232 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x59zn" event={"ID":"e8513867-3729-4d93-b8ca-45ecb69c50e6","Type":"ContainerDied","Data":"ec1634b3909c3b128dbf98f5658155ac371c891b6e1b1b64672b4b0ba678b266"} Mar 21 04:54:36 crc kubenswrapper[4775]: I0321 04:54:36.431797 4775 generic.go:334] "Generic (PLEG): container finished" podID="373fa4a4-80b8-4b32-a9ee-b272604d3adc" containerID="872ab07c719410279fd52976d7026dd1e51fa3df57a56a17fe5312e830ff9260" exitCode=0 Mar 21 04:54:36 crc kubenswrapper[4775]: I0321 04:54:36.431835 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gslhk" event={"ID":"373fa4a4-80b8-4b32-a9ee-b272604d3adc","Type":"ContainerDied","Data":"872ab07c719410279fd52976d7026dd1e51fa3df57a56a17fe5312e830ff9260"} Mar 21 04:54:37 crc kubenswrapper[4775]: I0321 04:54:37.438776 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gslhk" event={"ID":"373fa4a4-80b8-4b32-a9ee-b272604d3adc","Type":"ContainerStarted","Data":"e3764d781193778add49d893cc700b1f8d1f9c234b6c2effbeb58f7c8db40da8"} Mar 21 04:54:37 crc kubenswrapper[4775]: I0321 04:54:37.441160 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x59zn" event={"ID":"e8513867-3729-4d93-b8ca-45ecb69c50e6","Type":"ContainerStarted","Data":"9ad826a8c49b3029fe8ea5e2f130d732e68b23774780ea22f8580b36613c36b8"} Mar 21 04:54:37 crc kubenswrapper[4775]: I0321 04:54:37.460093 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gslhk" podStartSLOduration=3.02030458 podStartE2EDuration="5.460075211s" podCreationTimestamp="2026-03-21 04:54:32 +0000 UTC" firstStartedPulling="2026-03-21 04:54:34.409959712 +0000 UTC m=+427.386423336" lastFinishedPulling="2026-03-21 04:54:36.849730343 +0000 UTC m=+429.826193967" observedRunningTime="2026-03-21 04:54:37.459887556 +0000 UTC m=+430.436351180" watchObservedRunningTime="2026-03-21 04:54:37.460075211 +0000 UTC m=+430.436538845" Mar 21 04:54:37 crc kubenswrapper[4775]: I0321 04:54:37.486343 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x59zn" podStartSLOduration=2.413897692 podStartE2EDuration="4.486321169s" podCreationTimestamp="2026-03-21 04:54:33 +0000 UTC" firstStartedPulling="2026-03-21 04:54:34.415786357 +0000 UTC m=+427.392249981" lastFinishedPulling="2026-03-21 04:54:36.488209834 +0000 UTC m=+429.464673458" observedRunningTime="2026-03-21 04:54:37.474095194 +0000 UTC m=+430.450558828" watchObservedRunningTime="2026-03-21 04:54:37.486321169 +0000 UTC m=+430.462784793" Mar 21 04:54:40 crc kubenswrapper[4775]: I0321 04:54:40.798492 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-77cp9" Mar 21 04:54:40 crc kubenswrapper[4775]: I0321 04:54:40.798788 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-77cp9" Mar 21 04:54:40 crc kubenswrapper[4775]: I0321 04:54:40.847566 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-77cp9" Mar 21 04:54:41 crc kubenswrapper[4775]: I0321 04:54:41.370422 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h4wmk" Mar 21 04:54:41 crc kubenswrapper[4775]: I0321 04:54:41.370482 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h4wmk" Mar 21 04:54:41 crc kubenswrapper[4775]: I0321 04:54:41.411179 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h4wmk" Mar 21 04:54:41 crc kubenswrapper[4775]: I0321 04:54:41.498601 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-77cp9" Mar 21 04:54:41 crc kubenswrapper[4775]: I0321 04:54:41.498662 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h4wmk" Mar 21 04:54:43 crc kubenswrapper[4775]: I0321 04:54:43.175808 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gslhk" Mar 21 04:54:43 crc kubenswrapper[4775]: I0321 04:54:43.176855 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gslhk" Mar 21 04:54:43 crc kubenswrapper[4775]: I0321 04:54:43.217563 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gslhk" Mar 21 04:54:43 crc kubenswrapper[4775]: I0321 04:54:43.511416 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gslhk" Mar 21 04:54:43 crc kubenswrapper[4775]: I0321 04:54:43.759205 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x59zn" Mar 21 04:54:43 crc kubenswrapper[4775]: I0321 04:54:43.759269 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x59zn" Mar 21 04:54:43 crc kubenswrapper[4775]: I0321 04:54:43.795751 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x59zn" Mar 21 04:54:44 crc kubenswrapper[4775]: I0321 04:54:44.514360 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x59zn" Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.315244 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" podUID="60cd23b8-6e1c-492e-aeb5-8b16609d06d1" containerName="registry" containerID="cri-o://126ca17ac1331046f38ef3e95c39b37b4252bef89c1a0511f475a2f5ad46576e" gracePeriod=30 Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.519959 4775 generic.go:334] "Generic (PLEG): container finished" podID="60cd23b8-6e1c-492e-aeb5-8b16609d06d1" containerID="126ca17ac1331046f38ef3e95c39b37b4252bef89c1a0511f475a2f5ad46576e" exitCode=0 Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.520047 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" event={"ID":"60cd23b8-6e1c-492e-aeb5-8b16609d06d1","Type":"ContainerDied","Data":"126ca17ac1331046f38ef3e95c39b37b4252bef89c1a0511f475a2f5ad46576e"} Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.735068 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.773159 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-ca-trust-extracted\") pod \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.773320 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.773355 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-bound-sa-token\") pod \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.773380 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-registry-certificates\") pod \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.773401 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2j66\" (UniqueName: \"kubernetes.io/projected/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-kube-api-access-z2j66\") pod \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.773433 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-trusted-ca\") pod \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.773450 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-registry-tls\") pod \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.773470 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-installation-pull-secrets\") pod \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\" (UID: \"60cd23b8-6e1c-492e-aeb5-8b16609d06d1\") " Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.774368 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "60cd23b8-6e1c-492e-aeb5-8b16609d06d1" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.774982 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "60cd23b8-6e1c-492e-aeb5-8b16609d06d1" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.779047 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "60cd23b8-6e1c-492e-aeb5-8b16609d06d1" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.779280 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "60cd23b8-6e1c-492e-aeb5-8b16609d06d1" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.784193 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "60cd23b8-6e1c-492e-aeb5-8b16609d06d1" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.788948 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-kube-api-access-z2j66" (OuterVolumeSpecName: "kube-api-access-z2j66") pod "60cd23b8-6e1c-492e-aeb5-8b16609d06d1" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1"). InnerVolumeSpecName "kube-api-access-z2j66". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.793471 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "60cd23b8-6e1c-492e-aeb5-8b16609d06d1" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.794999 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "60cd23b8-6e1c-492e-aeb5-8b16609d06d1" (UID: "60cd23b8-6e1c-492e-aeb5-8b16609d06d1"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.874768 4775 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.874802 4775 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.874814 4775 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.874823 4775 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.874830 4775 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.874838 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2j66\" (UniqueName: \"kubernetes.io/projected/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-kube-api-access-z2j66\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:52 crc kubenswrapper[4775]: I0321 04:54:52.874846 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/60cd23b8-6e1c-492e-aeb5-8b16609d06d1-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 04:54:53 crc kubenswrapper[4775]: I0321 04:54:53.527509 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" event={"ID":"60cd23b8-6e1c-492e-aeb5-8b16609d06d1","Type":"ContainerDied","Data":"224b2bc46f2dabb71d4e0596c8a7cb1f7d3a4099e704fb5ab5882b52edcb5f8a"} Mar 21 04:54:53 crc kubenswrapper[4775]: I0321 04:54:53.527600 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hvd8t" Mar 21 04:54:53 crc kubenswrapper[4775]: I0321 04:54:53.527941 4775 scope.go:117] "RemoveContainer" containerID="126ca17ac1331046f38ef3e95c39b37b4252bef89c1a0511f475a2f5ad46576e" Mar 21 04:54:53 crc kubenswrapper[4775]: I0321 04:54:53.568260 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hvd8t"] Mar 21 04:54:53 crc kubenswrapper[4775]: I0321 04:54:53.574595 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hvd8t"] Mar 21 04:54:53 crc kubenswrapper[4775]: I0321 04:54:53.669311 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60cd23b8-6e1c-492e-aeb5-8b16609d06d1" path="/var/lib/kubelet/pods/60cd23b8-6e1c-492e-aeb5-8b16609d06d1/volumes" Mar 21 04:55:02 crc kubenswrapper[4775]: I0321 04:55:02.482707 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 04:55:02 crc kubenswrapper[4775]: I0321 04:55:02.483158 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 04:55:32 crc kubenswrapper[4775]: I0321 04:55:32.482808 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 04:55:32 crc kubenswrapper[4775]: I0321 04:55:32.483411 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 04:55:32 crc kubenswrapper[4775]: I0321 04:55:32.483460 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 04:55:32 crc kubenswrapper[4775]: I0321 04:55:32.483992 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bce245e457399f7eaba83f68d2114b3ba9e57f5a921fcddf334a4766f16c7398"} pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 04:55:32 crc kubenswrapper[4775]: I0321 04:55:32.484035 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" containerID="cri-o://bce245e457399f7eaba83f68d2114b3ba9e57f5a921fcddf334a4766f16c7398" gracePeriod=600 Mar 21 04:55:32 crc kubenswrapper[4775]: I0321 04:55:32.751040 4775 generic.go:334] "Generic (PLEG): container finished" podID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerID="bce245e457399f7eaba83f68d2114b3ba9e57f5a921fcddf334a4766f16c7398" exitCode=0 Mar 21 04:55:32 crc kubenswrapper[4775]: I0321 04:55:32.751168 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerDied","Data":"bce245e457399f7eaba83f68d2114b3ba9e57f5a921fcddf334a4766f16c7398"} Mar 21 04:55:32 crc kubenswrapper[4775]: I0321 04:55:32.751475 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerStarted","Data":"948421a752c90ac0f2fbc508e5358894c1d0ccf82922efe510f0505bdc2c2715"} Mar 21 04:55:32 crc kubenswrapper[4775]: I0321 04:55:32.751520 4775 scope.go:117] "RemoveContainer" containerID="94bb084d125f86725a85d0175e0edb9139a2dae93d13cf9dccf120d03cbeafee" Mar 21 04:56:00 crc kubenswrapper[4775]: I0321 04:56:00.131407 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567816-7wggn"] Mar 21 04:56:00 crc kubenswrapper[4775]: E0321 04:56:00.132194 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60cd23b8-6e1c-492e-aeb5-8b16609d06d1" containerName="registry" Mar 21 04:56:00 crc kubenswrapper[4775]: I0321 04:56:00.132211 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="60cd23b8-6e1c-492e-aeb5-8b16609d06d1" containerName="registry" Mar 21 04:56:00 crc kubenswrapper[4775]: I0321 04:56:00.132337 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="60cd23b8-6e1c-492e-aeb5-8b16609d06d1" containerName="registry" Mar 21 04:56:00 crc kubenswrapper[4775]: I0321 04:56:00.134499 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567816-7wggn" Mar 21 04:56:00 crc kubenswrapper[4775]: I0321 04:56:00.136793 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 04:56:00 crc kubenswrapper[4775]: I0321 04:56:00.136824 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 04:56:00 crc kubenswrapper[4775]: I0321 04:56:00.137329 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 04:56:00 crc kubenswrapper[4775]: I0321 04:56:00.145078 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567816-7wggn"] Mar 21 04:56:00 crc kubenswrapper[4775]: I0321 04:56:00.228850 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d4bx\" (UniqueName: \"kubernetes.io/projected/1edcd264-4604-481b-b146-1ed5a34badba-kube-api-access-9d4bx\") pod \"auto-csr-approver-29567816-7wggn\" (UID: \"1edcd264-4604-481b-b146-1ed5a34badba\") " pod="openshift-infra/auto-csr-approver-29567816-7wggn" Mar 21 04:56:00 crc kubenswrapper[4775]: I0321 04:56:00.329868 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d4bx\" (UniqueName: \"kubernetes.io/projected/1edcd264-4604-481b-b146-1ed5a34badba-kube-api-access-9d4bx\") pod \"auto-csr-approver-29567816-7wggn\" (UID: \"1edcd264-4604-481b-b146-1ed5a34badba\") " pod="openshift-infra/auto-csr-approver-29567816-7wggn" Mar 21 04:56:00 crc kubenswrapper[4775]: I0321 04:56:00.349017 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d4bx\" (UniqueName: \"kubernetes.io/projected/1edcd264-4604-481b-b146-1ed5a34badba-kube-api-access-9d4bx\") pod \"auto-csr-approver-29567816-7wggn\" (UID: \"1edcd264-4604-481b-b146-1ed5a34badba\") " pod="openshift-infra/auto-csr-approver-29567816-7wggn" Mar 21 04:56:00 crc kubenswrapper[4775]: I0321 04:56:00.455937 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567816-7wggn" Mar 21 04:56:00 crc kubenswrapper[4775]: I0321 04:56:00.824849 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567816-7wggn"] Mar 21 04:56:00 crc kubenswrapper[4775]: I0321 04:56:00.911266 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567816-7wggn" event={"ID":"1edcd264-4604-481b-b146-1ed5a34badba","Type":"ContainerStarted","Data":"f56a26e1863543e016e4efa5a83880872379157408d013285603de0a253952d7"} Mar 21 04:56:01 crc kubenswrapper[4775]: I0321 04:56:01.918730 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567816-7wggn" event={"ID":"1edcd264-4604-481b-b146-1ed5a34badba","Type":"ContainerStarted","Data":"67295240b136b0bcf5b3e83b034399ed927b48c5da3ccf9341e8f2fa7f873c3a"} Mar 21 04:56:01 crc kubenswrapper[4775]: I0321 04:56:01.930671 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29567816-7wggn" podStartSLOduration=1.154532562 podStartE2EDuration="1.930651583s" podCreationTimestamp="2026-03-21 04:56:00 +0000 UTC" firstStartedPulling="2026-03-21 04:56:00.836583609 +0000 UTC m=+513.813047233" lastFinishedPulling="2026-03-21 04:56:01.61270263 +0000 UTC m=+514.589166254" observedRunningTime="2026-03-21 04:56:01.929157044 +0000 UTC m=+514.905620688" watchObservedRunningTime="2026-03-21 04:56:01.930651583 +0000 UTC m=+514.907115207" Mar 21 04:56:02 crc kubenswrapper[4775]: E0321 04:56:02.129196 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1edcd264_4604_481b_b146_1ed5a34badba.slice/crio-67295240b136b0bcf5b3e83b034399ed927b48c5da3ccf9341e8f2fa7f873c3a.scope\": RecentStats: unable to find data in memory cache]" Mar 21 04:56:02 crc kubenswrapper[4775]: I0321 04:56:02.927521 4775 generic.go:334] "Generic (PLEG): container finished" podID="1edcd264-4604-481b-b146-1ed5a34badba" containerID="67295240b136b0bcf5b3e83b034399ed927b48c5da3ccf9341e8f2fa7f873c3a" exitCode=0 Mar 21 04:56:02 crc kubenswrapper[4775]: I0321 04:56:02.927590 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567816-7wggn" event={"ID":"1edcd264-4604-481b-b146-1ed5a34badba","Type":"ContainerDied","Data":"67295240b136b0bcf5b3e83b034399ed927b48c5da3ccf9341e8f2fa7f873c3a"} Mar 21 04:56:04 crc kubenswrapper[4775]: I0321 04:56:04.149649 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567816-7wggn" Mar 21 04:56:04 crc kubenswrapper[4775]: I0321 04:56:04.180186 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9d4bx\" (UniqueName: \"kubernetes.io/projected/1edcd264-4604-481b-b146-1ed5a34badba-kube-api-access-9d4bx\") pod \"1edcd264-4604-481b-b146-1ed5a34badba\" (UID: \"1edcd264-4604-481b-b146-1ed5a34badba\") " Mar 21 04:56:04 crc kubenswrapper[4775]: I0321 04:56:04.188955 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1edcd264-4604-481b-b146-1ed5a34badba-kube-api-access-9d4bx" (OuterVolumeSpecName: "kube-api-access-9d4bx") pod "1edcd264-4604-481b-b146-1ed5a34badba" (UID: "1edcd264-4604-481b-b146-1ed5a34badba"). InnerVolumeSpecName "kube-api-access-9d4bx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:56:04 crc kubenswrapper[4775]: I0321 04:56:04.281362 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9d4bx\" (UniqueName: \"kubernetes.io/projected/1edcd264-4604-481b-b146-1ed5a34badba-kube-api-access-9d4bx\") on node \"crc\" DevicePath \"\"" Mar 21 04:56:04 crc kubenswrapper[4775]: I0321 04:56:04.939571 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567816-7wggn" event={"ID":"1edcd264-4604-481b-b146-1ed5a34badba","Type":"ContainerDied","Data":"f56a26e1863543e016e4efa5a83880872379157408d013285603de0a253952d7"} Mar 21 04:56:04 crc kubenswrapper[4775]: I0321 04:56:04.939907 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f56a26e1863543e016e4efa5a83880872379157408d013285603de0a253952d7" Mar 21 04:56:04 crc kubenswrapper[4775]: I0321 04:56:04.939655 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567816-7wggn" Mar 21 04:56:04 crc kubenswrapper[4775]: I0321 04:56:04.985855 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567810-wb89g"] Mar 21 04:56:04 crc kubenswrapper[4775]: I0321 04:56:04.989902 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567810-wb89g"] Mar 21 04:56:05 crc kubenswrapper[4775]: I0321 04:56:05.670874 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd25c8a4-8047-4602-a95b-3308af65bd38" path="/var/lib/kubelet/pods/cd25c8a4-8047-4602-a95b-3308af65bd38/volumes" Mar 21 04:57:32 crc kubenswrapper[4775]: I0321 04:57:32.482745 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 04:57:32 crc kubenswrapper[4775]: I0321 04:57:32.483308 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 04:57:37 crc kubenswrapper[4775]: I0321 04:57:37.022965 4775 scope.go:117] "RemoveContainer" containerID="5c917ed3fbfe2c82bdf6a98ab031c5d0f484f2833ad0eb7cad0c0b48f3a9128d" Mar 21 04:58:00 crc kubenswrapper[4775]: I0321 04:58:00.144286 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567818-42hv5"] Mar 21 04:58:00 crc kubenswrapper[4775]: E0321 04:58:00.145164 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1edcd264-4604-481b-b146-1ed5a34badba" containerName="oc" Mar 21 04:58:00 crc kubenswrapper[4775]: I0321 04:58:00.145185 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1edcd264-4604-481b-b146-1ed5a34badba" containerName="oc" Mar 21 04:58:00 crc kubenswrapper[4775]: I0321 04:58:00.145354 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1edcd264-4604-481b-b146-1ed5a34badba" containerName="oc" Mar 21 04:58:00 crc kubenswrapper[4775]: I0321 04:58:00.145891 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567818-42hv5" Mar 21 04:58:00 crc kubenswrapper[4775]: I0321 04:58:00.148959 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 04:58:00 crc kubenswrapper[4775]: I0321 04:58:00.149226 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 04:58:00 crc kubenswrapper[4775]: I0321 04:58:00.149644 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 04:58:00 crc kubenswrapper[4775]: I0321 04:58:00.162184 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567818-42hv5"] Mar 21 04:58:00 crc kubenswrapper[4775]: I0321 04:58:00.320679 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b5mh\" (UniqueName: \"kubernetes.io/projected/40cf5461-a1ca-40cb-9ed3-9c2de90faad3-kube-api-access-4b5mh\") pod \"auto-csr-approver-29567818-42hv5\" (UID: \"40cf5461-a1ca-40cb-9ed3-9c2de90faad3\") " pod="openshift-infra/auto-csr-approver-29567818-42hv5" Mar 21 04:58:00 crc kubenswrapper[4775]: I0321 04:58:00.421717 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b5mh\" (UniqueName: \"kubernetes.io/projected/40cf5461-a1ca-40cb-9ed3-9c2de90faad3-kube-api-access-4b5mh\") pod \"auto-csr-approver-29567818-42hv5\" (UID: \"40cf5461-a1ca-40cb-9ed3-9c2de90faad3\") " pod="openshift-infra/auto-csr-approver-29567818-42hv5" Mar 21 04:58:00 crc kubenswrapper[4775]: I0321 04:58:00.445139 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b5mh\" (UniqueName: \"kubernetes.io/projected/40cf5461-a1ca-40cb-9ed3-9c2de90faad3-kube-api-access-4b5mh\") pod \"auto-csr-approver-29567818-42hv5\" (UID: \"40cf5461-a1ca-40cb-9ed3-9c2de90faad3\") " pod="openshift-infra/auto-csr-approver-29567818-42hv5" Mar 21 04:58:00 crc kubenswrapper[4775]: I0321 04:58:00.475889 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567818-42hv5" Mar 21 04:58:00 crc kubenswrapper[4775]: I0321 04:58:00.709708 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567818-42hv5"] Mar 21 04:58:00 crc kubenswrapper[4775]: I0321 04:58:00.720389 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 04:58:01 crc kubenswrapper[4775]: I0321 04:58:01.613237 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567818-42hv5" event={"ID":"40cf5461-a1ca-40cb-9ed3-9c2de90faad3","Type":"ContainerStarted","Data":"08556185bd2b965413587182184a1d1685909870272fa6ed0f264c756e900c6b"} Mar 21 04:58:02 crc kubenswrapper[4775]: I0321 04:58:02.482466 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 04:58:02 crc kubenswrapper[4775]: I0321 04:58:02.483066 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 04:58:02 crc kubenswrapper[4775]: I0321 04:58:02.624199 4775 generic.go:334] "Generic (PLEG): container finished" podID="40cf5461-a1ca-40cb-9ed3-9c2de90faad3" containerID="a6f778f857c6aa3ee7f297014df3be268b9de30de7297b84506cc38e86b9d649" exitCode=0 Mar 21 04:58:02 crc kubenswrapper[4775]: I0321 04:58:02.624276 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567818-42hv5" event={"ID":"40cf5461-a1ca-40cb-9ed3-9c2de90faad3","Type":"ContainerDied","Data":"a6f778f857c6aa3ee7f297014df3be268b9de30de7297b84506cc38e86b9d649"} Mar 21 04:58:03 crc kubenswrapper[4775]: I0321 04:58:03.852669 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567818-42hv5" Mar 21 04:58:03 crc kubenswrapper[4775]: I0321 04:58:03.967460 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4b5mh\" (UniqueName: \"kubernetes.io/projected/40cf5461-a1ca-40cb-9ed3-9c2de90faad3-kube-api-access-4b5mh\") pod \"40cf5461-a1ca-40cb-9ed3-9c2de90faad3\" (UID: \"40cf5461-a1ca-40cb-9ed3-9c2de90faad3\") " Mar 21 04:58:03 crc kubenswrapper[4775]: I0321 04:58:03.973955 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40cf5461-a1ca-40cb-9ed3-9c2de90faad3-kube-api-access-4b5mh" (OuterVolumeSpecName: "kube-api-access-4b5mh") pod "40cf5461-a1ca-40cb-9ed3-9c2de90faad3" (UID: "40cf5461-a1ca-40cb-9ed3-9c2de90faad3"). InnerVolumeSpecName "kube-api-access-4b5mh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 04:58:04 crc kubenswrapper[4775]: I0321 04:58:04.069112 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4b5mh\" (UniqueName: \"kubernetes.io/projected/40cf5461-a1ca-40cb-9ed3-9c2de90faad3-kube-api-access-4b5mh\") on node \"crc\" DevicePath \"\"" Mar 21 04:58:04 crc kubenswrapper[4775]: I0321 04:58:04.637289 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567818-42hv5" event={"ID":"40cf5461-a1ca-40cb-9ed3-9c2de90faad3","Type":"ContainerDied","Data":"08556185bd2b965413587182184a1d1685909870272fa6ed0f264c756e900c6b"} Mar 21 04:58:04 crc kubenswrapper[4775]: I0321 04:58:04.637345 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08556185bd2b965413587182184a1d1685909870272fa6ed0f264c756e900c6b" Mar 21 04:58:04 crc kubenswrapper[4775]: I0321 04:58:04.637365 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567818-42hv5" Mar 21 04:58:04 crc kubenswrapper[4775]: I0321 04:58:04.918053 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567812-6wppn"] Mar 21 04:58:04 crc kubenswrapper[4775]: I0321 04:58:04.922449 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567812-6wppn"] Mar 21 04:58:05 crc kubenswrapper[4775]: I0321 04:58:05.672292 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8101654-10fd-404b-ae0f-a098719418f4" path="/var/lib/kubelet/pods/d8101654-10fd-404b-ae0f-a098719418f4/volumes" Mar 21 04:58:32 crc kubenswrapper[4775]: I0321 04:58:32.482661 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 04:58:32 crc kubenswrapper[4775]: I0321 04:58:32.483289 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 04:58:32 crc kubenswrapper[4775]: I0321 04:58:32.483339 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 04:58:32 crc kubenswrapper[4775]: I0321 04:58:32.484024 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"948421a752c90ac0f2fbc508e5358894c1d0ccf82922efe510f0505bdc2c2715"} pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 04:58:32 crc kubenswrapper[4775]: I0321 04:58:32.484099 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" containerID="cri-o://948421a752c90ac0f2fbc508e5358894c1d0ccf82922efe510f0505bdc2c2715" gracePeriod=600 Mar 21 04:58:32 crc kubenswrapper[4775]: I0321 04:58:32.825800 4775 generic.go:334] "Generic (PLEG): container finished" podID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerID="948421a752c90ac0f2fbc508e5358894c1d0ccf82922efe510f0505bdc2c2715" exitCode=0 Mar 21 04:58:32 crc kubenswrapper[4775]: I0321 04:58:32.825901 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerDied","Data":"948421a752c90ac0f2fbc508e5358894c1d0ccf82922efe510f0505bdc2c2715"} Mar 21 04:58:32 crc kubenswrapper[4775]: I0321 04:58:32.826190 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerStarted","Data":"eaf646a6237d4b4ce6a8a82755505f01a336cf93fef407f02dbf82d68f5008b4"} Mar 21 04:58:32 crc kubenswrapper[4775]: I0321 04:58:32.826239 4775 scope.go:117] "RemoveContainer" containerID="bce245e457399f7eaba83f68d2114b3ba9e57f5a921fcddf334a4766f16c7398" Mar 21 04:58:37 crc kubenswrapper[4775]: I0321 04:58:37.075384 4775 scope.go:117] "RemoveContainer" containerID="4cd8fe98e0e79605364ea6850bbec4e960085c0c48458cb67fb164296fe16045" Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.731861 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-c2lsw"] Mar 21 04:59:51 crc kubenswrapper[4775]: E0321 04:59:51.733141 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40cf5461-a1ca-40cb-9ed3-9c2de90faad3" containerName="oc" Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.733160 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="40cf5461-a1ca-40cb-9ed3-9c2de90faad3" containerName="oc" Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.733300 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="40cf5461-a1ca-40cb-9ed3-9c2de90faad3" containerName="oc" Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.733940 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-c2lsw" Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.735407 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-ngckr"] Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.736381 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-ngckr" Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.736408 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.736591 4775 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-7lq2p" Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.736777 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.738854 4775 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-wdbmp" Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.745749 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsdtw\" (UniqueName: \"kubernetes.io/projected/4fdc8b75-b0a1-4ed3-9eee-6ee726dd0fbe-kube-api-access-zsdtw\") pod \"cert-manager-858654f9db-c2lsw\" (UID: \"4fdc8b75-b0a1-4ed3-9eee-6ee726dd0fbe\") " pod="cert-manager/cert-manager-858654f9db-c2lsw" Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.745805 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5njr\" (UniqueName: \"kubernetes.io/projected/172b2006-3394-469a-be7f-1b66d020fd45-kube-api-access-q5njr\") pod \"cert-manager-cainjector-cf98fcc89-ngckr\" (UID: \"172b2006-3394-469a-be7f-1b66d020fd45\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-ngckr" Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.757285 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-ngckr"] Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.762902 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-ms9hv"] Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.763599 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-ms9hv" Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.765876 4775 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-872m9" Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.788179 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-c2lsw"] Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.797697 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-ms9hv"] Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.847762 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsdtw\" (UniqueName: \"kubernetes.io/projected/4fdc8b75-b0a1-4ed3-9eee-6ee726dd0fbe-kube-api-access-zsdtw\") pod \"cert-manager-858654f9db-c2lsw\" (UID: \"4fdc8b75-b0a1-4ed3-9eee-6ee726dd0fbe\") " pod="cert-manager/cert-manager-858654f9db-c2lsw" Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.847877 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5njr\" (UniqueName: \"kubernetes.io/projected/172b2006-3394-469a-be7f-1b66d020fd45-kube-api-access-q5njr\") pod \"cert-manager-cainjector-cf98fcc89-ngckr\" (UID: \"172b2006-3394-469a-be7f-1b66d020fd45\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-ngckr" Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.847932 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42789\" (UniqueName: \"kubernetes.io/projected/670f734f-e215-441e-9b56-7251bc7f2484-kube-api-access-42789\") pod \"cert-manager-webhook-687f57d79b-ms9hv\" (UID: \"670f734f-e215-441e-9b56-7251bc7f2484\") " pod="cert-manager/cert-manager-webhook-687f57d79b-ms9hv" Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.873498 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5njr\" (UniqueName: \"kubernetes.io/projected/172b2006-3394-469a-be7f-1b66d020fd45-kube-api-access-q5njr\") pod \"cert-manager-cainjector-cf98fcc89-ngckr\" (UID: \"172b2006-3394-469a-be7f-1b66d020fd45\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-ngckr" Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.877291 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsdtw\" (UniqueName: \"kubernetes.io/projected/4fdc8b75-b0a1-4ed3-9eee-6ee726dd0fbe-kube-api-access-zsdtw\") pod \"cert-manager-858654f9db-c2lsw\" (UID: \"4fdc8b75-b0a1-4ed3-9eee-6ee726dd0fbe\") " pod="cert-manager/cert-manager-858654f9db-c2lsw" Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.949163 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42789\" (UniqueName: \"kubernetes.io/projected/670f734f-e215-441e-9b56-7251bc7f2484-kube-api-access-42789\") pod \"cert-manager-webhook-687f57d79b-ms9hv\" (UID: \"670f734f-e215-441e-9b56-7251bc7f2484\") " pod="cert-manager/cert-manager-webhook-687f57d79b-ms9hv" Mar 21 04:59:51 crc kubenswrapper[4775]: I0321 04:59:51.975919 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42789\" (UniqueName: \"kubernetes.io/projected/670f734f-e215-441e-9b56-7251bc7f2484-kube-api-access-42789\") pod \"cert-manager-webhook-687f57d79b-ms9hv\" (UID: \"670f734f-e215-441e-9b56-7251bc7f2484\") " pod="cert-manager/cert-manager-webhook-687f57d79b-ms9hv" Mar 21 04:59:52 crc kubenswrapper[4775]: I0321 04:59:52.055435 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-c2lsw" Mar 21 04:59:52 crc kubenswrapper[4775]: I0321 04:59:52.063245 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-ngckr" Mar 21 04:59:52 crc kubenswrapper[4775]: I0321 04:59:52.082437 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-ms9hv" Mar 21 04:59:52 crc kubenswrapper[4775]: I0321 04:59:52.312754 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-ngckr"] Mar 21 04:59:52 crc kubenswrapper[4775]: I0321 04:59:52.340848 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-ms9hv"] Mar 21 04:59:52 crc kubenswrapper[4775]: I0321 04:59:52.346056 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-ngckr" event={"ID":"172b2006-3394-469a-be7f-1b66d020fd45","Type":"ContainerStarted","Data":"19c5215685bfb668cb35aa2a26273b0b3c60e56f7ccd9949fe72bc98d0c2e2cf"} Mar 21 04:59:52 crc kubenswrapper[4775]: W0321 04:59:52.347774 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod670f734f_e215_441e_9b56_7251bc7f2484.slice/crio-2b091bba955a87c7e5271b0c1fcb64ec94db8632dd253448d1aeb06c1f732520 WatchSource:0}: Error finding container 2b091bba955a87c7e5271b0c1fcb64ec94db8632dd253448d1aeb06c1f732520: Status 404 returned error can't find the container with id 2b091bba955a87c7e5271b0c1fcb64ec94db8632dd253448d1aeb06c1f732520 Mar 21 04:59:52 crc kubenswrapper[4775]: I0321 04:59:52.370014 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-c2lsw"] Mar 21 04:59:52 crc kubenswrapper[4775]: W0321 04:59:52.378408 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4fdc8b75_b0a1_4ed3_9eee_6ee726dd0fbe.slice/crio-88e9bcf5f70d457cf4a7035b4c6a662a45207ada3ef6129d5a2d0c1f5d7644be WatchSource:0}: Error finding container 88e9bcf5f70d457cf4a7035b4c6a662a45207ada3ef6129d5a2d0c1f5d7644be: Status 404 returned error can't find the container with id 88e9bcf5f70d457cf4a7035b4c6a662a45207ada3ef6129d5a2d0c1f5d7644be Mar 21 04:59:53 crc kubenswrapper[4775]: I0321 04:59:53.356174 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-c2lsw" event={"ID":"4fdc8b75-b0a1-4ed3-9eee-6ee726dd0fbe","Type":"ContainerStarted","Data":"88e9bcf5f70d457cf4a7035b4c6a662a45207ada3ef6129d5a2d0c1f5d7644be"} Mar 21 04:59:53 crc kubenswrapper[4775]: I0321 04:59:53.357938 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-ms9hv" event={"ID":"670f734f-e215-441e-9b56-7251bc7f2484","Type":"ContainerStarted","Data":"2b091bba955a87c7e5271b0c1fcb64ec94db8632dd253448d1aeb06c1f732520"} Mar 21 04:59:56 crc kubenswrapper[4775]: I0321 04:59:56.378407 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-ngckr" event={"ID":"172b2006-3394-469a-be7f-1b66d020fd45","Type":"ContainerStarted","Data":"c70f86ffbfe854c6689cd59af95dda0da76c62bf835fda73a1e1b4aff21123e8"} Mar 21 04:59:56 crc kubenswrapper[4775]: I0321 04:59:56.380539 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-ms9hv" event={"ID":"670f734f-e215-441e-9b56-7251bc7f2484","Type":"ContainerStarted","Data":"ac927c4968781bf2bbd9aaf0fb4591462f558b972010214594ad0567776ff09b"} Mar 21 04:59:56 crc kubenswrapper[4775]: I0321 04:59:56.381372 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-ms9hv" Mar 21 04:59:56 crc kubenswrapper[4775]: I0321 04:59:56.383445 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-c2lsw" event={"ID":"4fdc8b75-b0a1-4ed3-9eee-6ee726dd0fbe","Type":"ContainerStarted","Data":"d1746e5e6672ac03f818cd2f4b17ba571d21e797dc6fbd87e03ac663b47fb65a"} Mar 21 04:59:56 crc kubenswrapper[4775]: I0321 04:59:56.400956 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-ngckr" podStartSLOduration=1.752043246 podStartE2EDuration="5.400927666s" podCreationTimestamp="2026-03-21 04:59:51 +0000 UTC" firstStartedPulling="2026-03-21 04:59:52.323302889 +0000 UTC m=+745.299766513" lastFinishedPulling="2026-03-21 04:59:55.972187299 +0000 UTC m=+748.948650933" observedRunningTime="2026-03-21 04:59:56.395838942 +0000 UTC m=+749.372302566" watchObservedRunningTime="2026-03-21 04:59:56.400927666 +0000 UTC m=+749.377391300" Mar 21 04:59:56 crc kubenswrapper[4775]: I0321 04:59:56.464525 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-c2lsw" podStartSLOduration=1.8094235859999999 podStartE2EDuration="5.464498971s" podCreationTimestamp="2026-03-21 04:59:51 +0000 UTC" firstStartedPulling="2026-03-21 04:59:52.381695828 +0000 UTC m=+745.358159452" lastFinishedPulling="2026-03-21 04:59:56.036771213 +0000 UTC m=+749.013234837" observedRunningTime="2026-03-21 04:59:56.458515272 +0000 UTC m=+749.434978926" watchObservedRunningTime="2026-03-21 04:59:56.464498971 +0000 UTC m=+749.440962605" Mar 21 04:59:56 crc kubenswrapper[4775]: I0321 04:59:56.489052 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-ms9hv" podStartSLOduration=1.818157943 podStartE2EDuration="5.489026864s" podCreationTimestamp="2026-03-21 04:59:51 +0000 UTC" firstStartedPulling="2026-03-21 04:59:52.354605923 +0000 UTC m=+745.331069547" lastFinishedPulling="2026-03-21 04:59:56.025474824 +0000 UTC m=+749.001938468" observedRunningTime="2026-03-21 04:59:56.486205604 +0000 UTC m=+749.462669238" watchObservedRunningTime="2026-03-21 04:59:56.489026864 +0000 UTC m=+749.465490498" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.158015 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567820-cngbx"] Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.159564 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567820-cngbx" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.162290 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.162480 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.162636 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.175288 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq"] Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.176469 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567820-cngbx"] Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.176606 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.210894 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.212628 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.217676 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/421d78dc-59bb-4b5b-9738-1eb6a6144b38-secret-volume\") pod \"collect-profiles-29567820-pcqhq\" (UID: \"421d78dc-59bb-4b5b-9738-1eb6a6144b38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.217752 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npphs\" (UniqueName: \"kubernetes.io/projected/1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3-kube-api-access-npphs\") pod \"auto-csr-approver-29567820-cngbx\" (UID: \"1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3\") " pod="openshift-infra/auto-csr-approver-29567820-cngbx" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.217826 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/421d78dc-59bb-4b5b-9738-1eb6a6144b38-config-volume\") pod \"collect-profiles-29567820-pcqhq\" (UID: \"421d78dc-59bb-4b5b-9738-1eb6a6144b38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.217852 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bk98\" (UniqueName: \"kubernetes.io/projected/421d78dc-59bb-4b5b-9738-1eb6a6144b38-kube-api-access-5bk98\") pod \"collect-profiles-29567820-pcqhq\" (UID: \"421d78dc-59bb-4b5b-9738-1eb6a6144b38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.228941 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq"] Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.319463 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/421d78dc-59bb-4b5b-9738-1eb6a6144b38-config-volume\") pod \"collect-profiles-29567820-pcqhq\" (UID: \"421d78dc-59bb-4b5b-9738-1eb6a6144b38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.319526 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bk98\" (UniqueName: \"kubernetes.io/projected/421d78dc-59bb-4b5b-9738-1eb6a6144b38-kube-api-access-5bk98\") pod \"collect-profiles-29567820-pcqhq\" (UID: \"421d78dc-59bb-4b5b-9738-1eb6a6144b38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.319564 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/421d78dc-59bb-4b5b-9738-1eb6a6144b38-secret-volume\") pod \"collect-profiles-29567820-pcqhq\" (UID: \"421d78dc-59bb-4b5b-9738-1eb6a6144b38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.319610 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npphs\" (UniqueName: \"kubernetes.io/projected/1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3-kube-api-access-npphs\") pod \"auto-csr-approver-29567820-cngbx\" (UID: \"1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3\") " pod="openshift-infra/auto-csr-approver-29567820-cngbx" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.321710 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/421d78dc-59bb-4b5b-9738-1eb6a6144b38-config-volume\") pod \"collect-profiles-29567820-pcqhq\" (UID: \"421d78dc-59bb-4b5b-9738-1eb6a6144b38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.327794 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/421d78dc-59bb-4b5b-9738-1eb6a6144b38-secret-volume\") pod \"collect-profiles-29567820-pcqhq\" (UID: \"421d78dc-59bb-4b5b-9738-1eb6a6144b38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.340170 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bk98\" (UniqueName: \"kubernetes.io/projected/421d78dc-59bb-4b5b-9738-1eb6a6144b38-kube-api-access-5bk98\") pod \"collect-profiles-29567820-pcqhq\" (UID: \"421d78dc-59bb-4b5b-9738-1eb6a6144b38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.342002 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npphs\" (UniqueName: \"kubernetes.io/projected/1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3-kube-api-access-npphs\") pod \"auto-csr-approver-29567820-cngbx\" (UID: \"1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3\") " pod="openshift-infra/auto-csr-approver-29567820-cngbx" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.527408 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567820-cngbx" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.541331 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq" Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.739558 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq"] Mar 21 05:00:00 crc kubenswrapper[4775]: W0321 05:00:00.747886 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod421d78dc_59bb_4b5b_9738_1eb6a6144b38.slice/crio-639089e1b2eb45331baee2fe75955c82e4844add60351411ed01a1c3e360ac7d WatchSource:0}: Error finding container 639089e1b2eb45331baee2fe75955c82e4844add60351411ed01a1c3e360ac7d: Status 404 returned error can't find the container with id 639089e1b2eb45331baee2fe75955c82e4844add60351411ed01a1c3e360ac7d Mar 21 05:00:00 crc kubenswrapper[4775]: I0321 05:00:00.775058 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567820-cngbx"] Mar 21 05:00:00 crc kubenswrapper[4775]: W0321 05:00:00.783276 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a4faca2_a8cb_41b5_a0cf_44be1cea8fd3.slice/crio-b2b9b761df7ee4ff2e31ace5255a68fec2fe5933b83b3c71b10847ca26c3bcde WatchSource:0}: Error finding container b2b9b761df7ee4ff2e31ace5255a68fec2fe5933b83b3c71b10847ca26c3bcde: Status 404 returned error can't find the container with id b2b9b761df7ee4ff2e31ace5255a68fec2fe5933b83b3c71b10847ca26c3bcde Mar 21 05:00:01 crc kubenswrapper[4775]: I0321 05:00:01.424184 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567820-cngbx" event={"ID":"1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3","Type":"ContainerStarted","Data":"b2b9b761df7ee4ff2e31ace5255a68fec2fe5933b83b3c71b10847ca26c3bcde"} Mar 21 05:00:01 crc kubenswrapper[4775]: I0321 05:00:01.426254 4775 generic.go:334] "Generic (PLEG): container finished" podID="421d78dc-59bb-4b5b-9738-1eb6a6144b38" containerID="19febf6fd2202a1482d38e4d840812756bb31244e9ab9826c5c08685c9d1d70a" exitCode=0 Mar 21 05:00:01 crc kubenswrapper[4775]: I0321 05:00:01.426284 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq" event={"ID":"421d78dc-59bb-4b5b-9738-1eb6a6144b38","Type":"ContainerDied","Data":"19febf6fd2202a1482d38e4d840812756bb31244e9ab9826c5c08685c9d1d70a"} Mar 21 05:00:01 crc kubenswrapper[4775]: I0321 05:00:01.426302 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq" event={"ID":"421d78dc-59bb-4b5b-9738-1eb6a6144b38","Type":"ContainerStarted","Data":"639089e1b2eb45331baee2fe75955c82e4844add60351411ed01a1c3e360ac7d"} Mar 21 05:00:01 crc kubenswrapper[4775]: I0321 05:00:01.683282 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mzqtk"] Mar 21 05:00:01 crc kubenswrapper[4775]: I0321 05:00:01.683720 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovn-controller" containerID="cri-o://5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde" gracePeriod=30 Mar 21 05:00:01 crc kubenswrapper[4775]: I0321 05:00:01.684108 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="sbdb" containerID="cri-o://2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d" gracePeriod=30 Mar 21 05:00:01 crc kubenswrapper[4775]: I0321 05:00:01.684183 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="nbdb" containerID="cri-o://5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99" gracePeriod=30 Mar 21 05:00:01 crc kubenswrapper[4775]: I0321 05:00:01.684223 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="northd" containerID="cri-o://f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28" gracePeriod=30 Mar 21 05:00:01 crc kubenswrapper[4775]: I0321 05:00:01.684261 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4" gracePeriod=30 Mar 21 05:00:01 crc kubenswrapper[4775]: I0321 05:00:01.684304 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="kube-rbac-proxy-node" containerID="cri-o://6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc" gracePeriod=30 Mar 21 05:00:01 crc kubenswrapper[4775]: I0321 05:00:01.684342 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovn-acl-logging" containerID="cri-o://6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c" gracePeriod=30 Mar 21 05:00:01 crc kubenswrapper[4775]: I0321 05:00:01.734949 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovnkube-controller" containerID="cri-o://294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646" gracePeriod=30 Mar 21 05:00:01 crc kubenswrapper[4775]: I0321 05:00:01.998234 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzqtk_a69d31f5-deeb-4860-be96-ed5547831685/ovnkube-controller/3.log" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.000854 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzqtk_a69d31f5-deeb-4860-be96-ed5547831685/ovn-acl-logging/0.log" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.001467 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzqtk_a69d31f5-deeb-4860-be96-ed5547831685/ovn-controller/0.log" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.001995 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.039768 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-run-systemd\") pod \"a69d31f5-deeb-4860-be96-ed5547831685\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.039835 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-log-socket\") pod \"a69d31f5-deeb-4860-be96-ed5547831685\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.039881 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a69d31f5-deeb-4860-be96-ed5547831685-ovnkube-config\") pod \"a69d31f5-deeb-4860-be96-ed5547831685\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.039961 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-var-lib-openvswitch\") pod \"a69d31f5-deeb-4860-be96-ed5547831685\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.039996 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-run-netns\") pod \"a69d31f5-deeb-4860-be96-ed5547831685\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.040018 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-var-lib-cni-networks-ovn-kubernetes\") pod \"a69d31f5-deeb-4860-be96-ed5547831685\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.040040 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a69d31f5-deeb-4860-be96-ed5547831685-env-overrides\") pod \"a69d31f5-deeb-4860-be96-ed5547831685\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.040060 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a69d31f5-deeb-4860-be96-ed5547831685-ovn-node-metrics-cert\") pod \"a69d31f5-deeb-4860-be96-ed5547831685\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.040083 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-slash\") pod \"a69d31f5-deeb-4860-be96-ed5547831685\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.040106 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-kubelet\") pod \"a69d31f5-deeb-4860-be96-ed5547831685\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.040159 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-etc-openvswitch\") pod \"a69d31f5-deeb-4860-be96-ed5547831685\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.040179 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-node-log\") pod \"a69d31f5-deeb-4860-be96-ed5547831685\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.040318 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "a69d31f5-deeb-4860-be96-ed5547831685" (UID: "a69d31f5-deeb-4860-be96-ed5547831685"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.040356 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-slash" (OuterVolumeSpecName: "host-slash") pod "a69d31f5-deeb-4860-be96-ed5547831685" (UID: "a69d31f5-deeb-4860-be96-ed5547831685"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.040595 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "a69d31f5-deeb-4860-be96-ed5547831685" (UID: "a69d31f5-deeb-4860-be96-ed5547831685"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.040604 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "a69d31f5-deeb-4860-be96-ed5547831685" (UID: "a69d31f5-deeb-4860-be96-ed5547831685"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.040655 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "a69d31f5-deeb-4860-be96-ed5547831685" (UID: "a69d31f5-deeb-4860-be96-ed5547831685"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.040632 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-log-socket" (OuterVolumeSpecName: "log-socket") pod "a69d31f5-deeb-4860-be96-ed5547831685" (UID: "a69d31f5-deeb-4860-be96-ed5547831685"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.040727 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "a69d31f5-deeb-4860-be96-ed5547831685" (UID: "a69d31f5-deeb-4860-be96-ed5547831685"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.040821 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a69d31f5-deeb-4860-be96-ed5547831685-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "a69d31f5-deeb-4860-be96-ed5547831685" (UID: "a69d31f5-deeb-4860-be96-ed5547831685"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.040848 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-node-log" (OuterVolumeSpecName: "node-log") pod "a69d31f5-deeb-4860-be96-ed5547831685" (UID: "a69d31f5-deeb-4860-be96-ed5547831685"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.041168 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a69d31f5-deeb-4860-be96-ed5547831685-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "a69d31f5-deeb-4860-be96-ed5547831685" (UID: "a69d31f5-deeb-4860-be96-ed5547831685"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.041229 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-run-ovn\") pod \"a69d31f5-deeb-4860-be96-ed5547831685\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.041275 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-run-ovn-kubernetes\") pod \"a69d31f5-deeb-4860-be96-ed5547831685\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.041298 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-cni-bin\") pod \"a69d31f5-deeb-4860-be96-ed5547831685\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.041327 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-cni-netd\") pod \"a69d31f5-deeb-4860-be96-ed5547831685\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.041279 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "a69d31f5-deeb-4860-be96-ed5547831685" (UID: "a69d31f5-deeb-4860-be96-ed5547831685"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.041346 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-systemd-units\") pod \"a69d31f5-deeb-4860-be96-ed5547831685\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.041362 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "a69d31f5-deeb-4860-be96-ed5547831685" (UID: "a69d31f5-deeb-4860-be96-ed5547831685"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.041395 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "a69d31f5-deeb-4860-be96-ed5547831685" (UID: "a69d31f5-deeb-4860-be96-ed5547831685"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.041374 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "a69d31f5-deeb-4860-be96-ed5547831685" (UID: "a69d31f5-deeb-4860-be96-ed5547831685"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.041426 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "a69d31f5-deeb-4860-be96-ed5547831685" (UID: "a69d31f5-deeb-4860-be96-ed5547831685"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.041452 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-run-openvswitch\") pod \"a69d31f5-deeb-4860-be96-ed5547831685\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.041494 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7w6b\" (UniqueName: \"kubernetes.io/projected/a69d31f5-deeb-4860-be96-ed5547831685-kube-api-access-h7w6b\") pod \"a69d31f5-deeb-4860-be96-ed5547831685\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.041524 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "a69d31f5-deeb-4860-be96-ed5547831685" (UID: "a69d31f5-deeb-4860-be96-ed5547831685"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.041556 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a69d31f5-deeb-4860-be96-ed5547831685-ovnkube-script-lib\") pod \"a69d31f5-deeb-4860-be96-ed5547831685\" (UID: \"a69d31f5-deeb-4860-be96-ed5547831685\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.042022 4775 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-log-socket\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.042057 4775 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a69d31f5-deeb-4860-be96-ed5547831685-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.042205 4775 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.042304 4775 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.042361 4775 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.042379 4775 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a69d31f5-deeb-4860-be96-ed5547831685-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.042397 4775 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-slash\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.042413 4775 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.042429 4775 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.042445 4775 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-node-log\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.042459 4775 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.042474 4775 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.042492 4775 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.042507 4775 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.042521 4775 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.042536 4775 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.042609 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a69d31f5-deeb-4860-be96-ed5547831685-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "a69d31f5-deeb-4860-be96-ed5547831685" (UID: "a69d31f5-deeb-4860-be96-ed5547831685"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.046009 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a69d31f5-deeb-4860-be96-ed5547831685-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "a69d31f5-deeb-4860-be96-ed5547831685" (UID: "a69d31f5-deeb-4860-be96-ed5547831685"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.046640 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a69d31f5-deeb-4860-be96-ed5547831685-kube-api-access-h7w6b" (OuterVolumeSpecName: "kube-api-access-h7w6b") pod "a69d31f5-deeb-4860-be96-ed5547831685" (UID: "a69d31f5-deeb-4860-be96-ed5547831685"). InnerVolumeSpecName "kube-api-access-h7w6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.065822 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-sbftc"] Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.066471 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="kubecfg-setup" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.066497 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="kubecfg-setup" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.066518 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovnkube-controller" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.066541 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovnkube-controller" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.066550 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="kube-rbac-proxy-ovn-metrics" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.066560 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="kube-rbac-proxy-ovn-metrics" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.066576 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovn-controller" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.066595 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovn-controller" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.066611 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="nbdb" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.066621 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="nbdb" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.066643 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovnkube-controller" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.066656 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovnkube-controller" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.066666 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovnkube-controller" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.066674 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovnkube-controller" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.066703 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovnkube-controller" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.066713 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovnkube-controller" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.066727 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovn-acl-logging" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.066735 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovn-acl-logging" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.066750 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="kube-rbac-proxy-node" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.066759 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="kube-rbac-proxy-node" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.066771 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="northd" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.066779 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="northd" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.066793 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="sbdb" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.066801 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="sbdb" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.066942 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="nbdb" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.066958 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="kube-rbac-proxy-ovn-metrics" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.066970 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovnkube-controller" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.066983 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="sbdb" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.066996 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovnkube-controller" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.067006 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovnkube-controller" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.067015 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="kube-rbac-proxy-node" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.067028 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovn-acl-logging" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.067039 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovnkube-controller" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.067048 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="northd" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.067057 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovn-controller" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.067231 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovnkube-controller" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.067242 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovnkube-controller" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.067368 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a69d31f5-deeb-4860-be96-ed5547831685" containerName="ovnkube-controller" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.069408 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "a69d31f5-deeb-4860-be96-ed5547831685" (UID: "a69d31f5-deeb-4860-be96-ed5547831685"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.069580 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.085240 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-ms9hv" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.144103 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rbww\" (UniqueName: \"kubernetes.io/projected/82b41c52-e915-4cf3-808e-79211ba01fce-kube-api-access-5rbww\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.144512 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-systemd-units\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.144673 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-var-lib-openvswitch\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.144778 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-slash\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.144870 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-run-ovn\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.144967 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.145064 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-run-netns\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.145180 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-kubelet\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.145297 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/82b41c52-e915-4cf3-808e-79211ba01fce-ovnkube-config\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.145440 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-cni-bin\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.145590 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/82b41c52-e915-4cf3-808e-79211ba01fce-env-overrides\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.145743 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/82b41c52-e915-4cf3-808e-79211ba01fce-ovn-node-metrics-cert\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.145884 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-etc-openvswitch\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.146029 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-run-ovn-kubernetes\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.146209 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-run-systemd\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.146368 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/82b41c52-e915-4cf3-808e-79211ba01fce-ovnkube-script-lib\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.146507 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-node-log\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.146642 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-run-openvswitch\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.146780 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-cni-netd\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.146925 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-log-socket\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.147221 4775 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a69d31f5-deeb-4860-be96-ed5547831685-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.147321 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7w6b\" (UniqueName: \"kubernetes.io/projected/a69d31f5-deeb-4860-be96-ed5547831685-kube-api-access-h7w6b\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.147411 4775 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a69d31f5-deeb-4860-be96-ed5547831685-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.147482 4775 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a69d31f5-deeb-4860-be96-ed5547831685-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.248394 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-run-openvswitch\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.248434 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-cni-netd\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.248455 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-log-socket\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.248484 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rbww\" (UniqueName: \"kubernetes.io/projected/82b41c52-e915-4cf3-808e-79211ba01fce-kube-api-access-5rbww\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.248504 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-systemd-units\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.248522 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-var-lib-openvswitch\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.248521 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-run-openvswitch\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.248588 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-slash\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.248541 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-slash\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.248630 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-log-socket\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.248644 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-cni-netd\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.248879 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-run-ovn\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.248917 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.248939 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-var-lib-openvswitch\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.248960 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-systemd-units\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.248964 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-run-netns\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.248990 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-run-ovn\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.249000 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-kubelet\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.249016 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.249043 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-run-netns\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.249035 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/82b41c52-e915-4cf3-808e-79211ba01fce-ovnkube-config\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.249064 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-kubelet\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.249157 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-cni-bin\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.249224 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/82b41c52-e915-4cf3-808e-79211ba01fce-env-overrides\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.249254 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/82b41c52-e915-4cf3-808e-79211ba01fce-ovn-node-metrics-cert\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.249292 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-etc-openvswitch\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.249317 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-run-ovn-kubernetes\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.249327 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-cni-bin\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.249391 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-etc-openvswitch\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.249400 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-run-systemd\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.249430 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-run-systemd\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.249473 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-host-run-ovn-kubernetes\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.249490 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/82b41c52-e915-4cf3-808e-79211ba01fce-ovnkube-script-lib\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.249530 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-node-log\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.249625 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/82b41c52-e915-4cf3-808e-79211ba01fce-node-log\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.249725 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/82b41c52-e915-4cf3-808e-79211ba01fce-ovnkube-config\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.250169 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/82b41c52-e915-4cf3-808e-79211ba01fce-env-overrides\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.250656 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/82b41c52-e915-4cf3-808e-79211ba01fce-ovnkube-script-lib\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.255078 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/82b41c52-e915-4cf3-808e-79211ba01fce-ovn-node-metrics-cert\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.272318 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rbww\" (UniqueName: \"kubernetes.io/projected/82b41c52-e915-4cf3-808e-79211ba01fce-kube-api-access-5rbww\") pod \"ovnkube-node-sbftc\" (UID: \"82b41c52-e915-4cf3-808e-79211ba01fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.386633 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.437098 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzqtk_a69d31f5-deeb-4860-be96-ed5547831685/ovnkube-controller/3.log" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.439810 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzqtk_a69d31f5-deeb-4860-be96-ed5547831685/ovn-acl-logging/0.log" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.440621 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzqtk_a69d31f5-deeb-4860-be96-ed5547831685/ovn-controller/0.log" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441173 4775 generic.go:334] "Generic (PLEG): container finished" podID="a69d31f5-deeb-4860-be96-ed5547831685" containerID="294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646" exitCode=0 Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441214 4775 generic.go:334] "Generic (PLEG): container finished" podID="a69d31f5-deeb-4860-be96-ed5547831685" containerID="2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d" exitCode=0 Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441231 4775 generic.go:334] "Generic (PLEG): container finished" podID="a69d31f5-deeb-4860-be96-ed5547831685" containerID="5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99" exitCode=0 Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441250 4775 generic.go:334] "Generic (PLEG): container finished" podID="a69d31f5-deeb-4860-be96-ed5547831685" containerID="f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28" exitCode=0 Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441266 4775 generic.go:334] "Generic (PLEG): container finished" podID="a69d31f5-deeb-4860-be96-ed5547831685" containerID="70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4" exitCode=0 Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441279 4775 generic.go:334] "Generic (PLEG): container finished" podID="a69d31f5-deeb-4860-be96-ed5547831685" containerID="6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc" exitCode=0 Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441294 4775 generic.go:334] "Generic (PLEG): container finished" podID="a69d31f5-deeb-4860-be96-ed5547831685" containerID="6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c" exitCode=143 Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441311 4775 generic.go:334] "Generic (PLEG): container finished" podID="a69d31f5-deeb-4860-be96-ed5547831685" containerID="5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde" exitCode=143 Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441387 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerDied","Data":"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441431 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerDied","Data":"2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441454 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerDied","Data":"5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441473 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerDied","Data":"f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441493 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerDied","Data":"70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441505 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441527 4775 scope.go:117] "RemoveContainer" containerID="294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441512 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerDied","Data":"6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441743 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441760 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441772 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441787 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441801 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441816 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441829 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441844 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441859 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441882 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerDied","Data":"6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441907 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441923 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441937 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441953 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441966 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441979 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.441993 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442006 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442020 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442035 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442055 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerDied","Data":"5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442078 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442094 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442109 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442206 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442218 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442229 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442240 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442251 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442262 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442273 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442288 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzqtk" event={"ID":"a69d31f5-deeb-4860-be96-ed5547831685","Type":"ContainerDied","Data":"a053283b74d1089f3887c9fbc93ee3063db45e363837d9c9231a5cf45771e07b"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442306 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442320 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442331 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442342 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442353 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442364 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442375 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442387 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442398 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.442409 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.443723 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" event={"ID":"82b41c52-e915-4cf3-808e-79211ba01fce","Type":"ContainerStarted","Data":"b1c177be1dcd02031cb1c2daefe3b54226235529cfff4a15959f6d5ad20824e0"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.447182 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-556rg_e77ec218-42da-4f07-b214-184c4f3b20f3/kube-multus/2.log" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.448162 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-556rg_e77ec218-42da-4f07-b214-184c4f3b20f3/kube-multus/1.log" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.448219 4775 generic.go:334] "Generic (PLEG): container finished" podID="e77ec218-42da-4f07-b214-184c4f3b20f3" containerID="ad12949f26afe1756d4a6c0d01069cb26a8928bb36cc24602d4ab1bbde117f9e" exitCode=2 Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.448373 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-556rg" event={"ID":"e77ec218-42da-4f07-b214-184c4f3b20f3","Type":"ContainerDied","Data":"ad12949f26afe1756d4a6c0d01069cb26a8928bb36cc24602d4ab1bbde117f9e"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.448483 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eaf83cd6f4a806265a871a5672f1a6939180308c2815b87ca254164f26d05aae"} Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.449103 4775 scope.go:117] "RemoveContainer" containerID="ad12949f26afe1756d4a6c0d01069cb26a8928bb36cc24602d4ab1bbde117f9e" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.449395 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-556rg_openshift-multus(e77ec218-42da-4f07-b214-184c4f3b20f3)\"" pod="openshift-multus/multus-556rg" podUID="e77ec218-42da-4f07-b214-184c4f3b20f3" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.480965 4775 scope.go:117] "RemoveContainer" containerID="7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.536851 4775 scope.go:117] "RemoveContainer" containerID="2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.537234 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.546493 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mzqtk"] Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.550761 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mzqtk"] Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.552381 4775 scope.go:117] "RemoveContainer" containerID="5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.572404 4775 scope.go:117] "RemoveContainer" containerID="f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.590373 4775 scope.go:117] "RemoveContainer" containerID="70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.653492 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/421d78dc-59bb-4b5b-9738-1eb6a6144b38-secret-volume\") pod \"421d78dc-59bb-4b5b-9738-1eb6a6144b38\" (UID: \"421d78dc-59bb-4b5b-9738-1eb6a6144b38\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.653645 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/421d78dc-59bb-4b5b-9738-1eb6a6144b38-config-volume\") pod \"421d78dc-59bb-4b5b-9738-1eb6a6144b38\" (UID: \"421d78dc-59bb-4b5b-9738-1eb6a6144b38\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.653706 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bk98\" (UniqueName: \"kubernetes.io/projected/421d78dc-59bb-4b5b-9738-1eb6a6144b38-kube-api-access-5bk98\") pod \"421d78dc-59bb-4b5b-9738-1eb6a6144b38\" (UID: \"421d78dc-59bb-4b5b-9738-1eb6a6144b38\") " Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.654398 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/421d78dc-59bb-4b5b-9738-1eb6a6144b38-config-volume" (OuterVolumeSpecName: "config-volume") pod "421d78dc-59bb-4b5b-9738-1eb6a6144b38" (UID: "421d78dc-59bb-4b5b-9738-1eb6a6144b38"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.657285 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/421d78dc-59bb-4b5b-9738-1eb6a6144b38-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "421d78dc-59bb-4b5b-9738-1eb6a6144b38" (UID: "421d78dc-59bb-4b5b-9738-1eb6a6144b38"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.657951 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/421d78dc-59bb-4b5b-9738-1eb6a6144b38-kube-api-access-5bk98" (OuterVolumeSpecName: "kube-api-access-5bk98") pod "421d78dc-59bb-4b5b-9738-1eb6a6144b38" (UID: "421d78dc-59bb-4b5b-9738-1eb6a6144b38"). InnerVolumeSpecName "kube-api-access-5bk98". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.658638 4775 scope.go:117] "RemoveContainer" containerID="6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.673679 4775 scope.go:117] "RemoveContainer" containerID="6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.689786 4775 scope.go:117] "RemoveContainer" containerID="5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.701908 4775 scope.go:117] "RemoveContainer" containerID="effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.716739 4775 scope.go:117] "RemoveContainer" containerID="294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.717276 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646\": container with ID starting with 294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646 not found: ID does not exist" containerID="294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.717324 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646"} err="failed to get container status \"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646\": rpc error: code = NotFound desc = could not find container \"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646\": container with ID starting with 294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.717352 4775 scope.go:117] "RemoveContainer" containerID="7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.717675 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a\": container with ID starting with 7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a not found: ID does not exist" containerID="7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.717718 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a"} err="failed to get container status \"7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a\": rpc error: code = NotFound desc = could not find container \"7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a\": container with ID starting with 7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.717746 4775 scope.go:117] "RemoveContainer" containerID="2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.718030 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\": container with ID starting with 2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d not found: ID does not exist" containerID="2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.718059 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d"} err="failed to get container status \"2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\": rpc error: code = NotFound desc = could not find container \"2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\": container with ID starting with 2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.718073 4775 scope.go:117] "RemoveContainer" containerID="5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.718517 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\": container with ID starting with 5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99 not found: ID does not exist" containerID="5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.718544 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99"} err="failed to get container status \"5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\": rpc error: code = NotFound desc = could not find container \"5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\": container with ID starting with 5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.718564 4775 scope.go:117] "RemoveContainer" containerID="f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.718820 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\": container with ID starting with f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28 not found: ID does not exist" containerID="f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.718846 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28"} err="failed to get container status \"f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\": rpc error: code = NotFound desc = could not find container \"f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\": container with ID starting with f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.718859 4775 scope.go:117] "RemoveContainer" containerID="70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.719091 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\": container with ID starting with 70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4 not found: ID does not exist" containerID="70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.719109 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4"} err="failed to get container status \"70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\": rpc error: code = NotFound desc = could not find container \"70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\": container with ID starting with 70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.719143 4775 scope.go:117] "RemoveContainer" containerID="6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.719492 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\": container with ID starting with 6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc not found: ID does not exist" containerID="6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.719512 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc"} err="failed to get container status \"6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\": rpc error: code = NotFound desc = could not find container \"6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\": container with ID starting with 6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.719525 4775 scope.go:117] "RemoveContainer" containerID="6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.719734 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\": container with ID starting with 6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c not found: ID does not exist" containerID="6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.719758 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c"} err="failed to get container status \"6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\": rpc error: code = NotFound desc = could not find container \"6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\": container with ID starting with 6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.719771 4775 scope.go:117] "RemoveContainer" containerID="5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.720004 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\": container with ID starting with 5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde not found: ID does not exist" containerID="5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.720025 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde"} err="failed to get container status \"5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\": rpc error: code = NotFound desc = could not find container \"5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\": container with ID starting with 5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.720041 4775 scope.go:117] "RemoveContainer" containerID="effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6" Mar 21 05:00:02 crc kubenswrapper[4775]: E0321 05:00:02.720299 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\": container with ID starting with effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6 not found: ID does not exist" containerID="effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.720331 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6"} err="failed to get container status \"effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\": rpc error: code = NotFound desc = could not find container \"effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\": container with ID starting with effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.720349 4775 scope.go:117] "RemoveContainer" containerID="294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.720564 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646"} err="failed to get container status \"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646\": rpc error: code = NotFound desc = could not find container \"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646\": container with ID starting with 294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.720587 4775 scope.go:117] "RemoveContainer" containerID="7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.720767 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a"} err="failed to get container status \"7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a\": rpc error: code = NotFound desc = could not find container \"7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a\": container with ID starting with 7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.720788 4775 scope.go:117] "RemoveContainer" containerID="2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.721054 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d"} err="failed to get container status \"2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\": rpc error: code = NotFound desc = could not find container \"2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\": container with ID starting with 2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.721094 4775 scope.go:117] "RemoveContainer" containerID="5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.721335 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99"} err="failed to get container status \"5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\": rpc error: code = NotFound desc = could not find container \"5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\": container with ID starting with 5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.721355 4775 scope.go:117] "RemoveContainer" containerID="f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.721600 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28"} err="failed to get container status \"f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\": rpc error: code = NotFound desc = could not find container \"f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\": container with ID starting with f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.721627 4775 scope.go:117] "RemoveContainer" containerID="70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.721870 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4"} err="failed to get container status \"70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\": rpc error: code = NotFound desc = could not find container \"70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\": container with ID starting with 70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.721911 4775 scope.go:117] "RemoveContainer" containerID="6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.722374 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc"} err="failed to get container status \"6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\": rpc error: code = NotFound desc = could not find container \"6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\": container with ID starting with 6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.722397 4775 scope.go:117] "RemoveContainer" containerID="6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.722668 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c"} err="failed to get container status \"6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\": rpc error: code = NotFound desc = could not find container \"6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\": container with ID starting with 6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.722692 4775 scope.go:117] "RemoveContainer" containerID="5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.722874 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde"} err="failed to get container status \"5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\": rpc error: code = NotFound desc = could not find container \"5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\": container with ID starting with 5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.722894 4775 scope.go:117] "RemoveContainer" containerID="effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.723101 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6"} err="failed to get container status \"effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\": rpc error: code = NotFound desc = could not find container \"effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\": container with ID starting with effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.723135 4775 scope.go:117] "RemoveContainer" containerID="294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.723375 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646"} err="failed to get container status \"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646\": rpc error: code = NotFound desc = could not find container \"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646\": container with ID starting with 294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.723395 4775 scope.go:117] "RemoveContainer" containerID="7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.723687 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a"} err="failed to get container status \"7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a\": rpc error: code = NotFound desc = could not find container \"7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a\": container with ID starting with 7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.723729 4775 scope.go:117] "RemoveContainer" containerID="2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.723961 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d"} err="failed to get container status \"2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\": rpc error: code = NotFound desc = could not find container \"2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\": container with ID starting with 2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.723983 4775 scope.go:117] "RemoveContainer" containerID="5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.724195 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99"} err="failed to get container status \"5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\": rpc error: code = NotFound desc = could not find container \"5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\": container with ID starting with 5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.724212 4775 scope.go:117] "RemoveContainer" containerID="f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.724462 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28"} err="failed to get container status \"f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\": rpc error: code = NotFound desc = could not find container \"f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\": container with ID starting with f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.724511 4775 scope.go:117] "RemoveContainer" containerID="70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.724731 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4"} err="failed to get container status \"70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\": rpc error: code = NotFound desc = could not find container \"70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\": container with ID starting with 70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.724750 4775 scope.go:117] "RemoveContainer" containerID="6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.724973 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc"} err="failed to get container status \"6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\": rpc error: code = NotFound desc = could not find container \"6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\": container with ID starting with 6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.724993 4775 scope.go:117] "RemoveContainer" containerID="6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.725195 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c"} err="failed to get container status \"6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\": rpc error: code = NotFound desc = could not find container \"6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\": container with ID starting with 6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.725214 4775 scope.go:117] "RemoveContainer" containerID="5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.725440 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde"} err="failed to get container status \"5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\": rpc error: code = NotFound desc = could not find container \"5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\": container with ID starting with 5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.725464 4775 scope.go:117] "RemoveContainer" containerID="effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.725745 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6"} err="failed to get container status \"effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\": rpc error: code = NotFound desc = could not find container \"effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\": container with ID starting with effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.725783 4775 scope.go:117] "RemoveContainer" containerID="294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.726030 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646"} err="failed to get container status \"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646\": rpc error: code = NotFound desc = could not find container \"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646\": container with ID starting with 294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.726059 4775 scope.go:117] "RemoveContainer" containerID="7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.726307 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a"} err="failed to get container status \"7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a\": rpc error: code = NotFound desc = could not find container \"7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a\": container with ID starting with 7d306fe24fb21d9e53a0846533716c9f52156736305200511b0df65cf5c69a6a not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.726335 4775 scope.go:117] "RemoveContainer" containerID="2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.726607 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d"} err="failed to get container status \"2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\": rpc error: code = NotFound desc = could not find container \"2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d\": container with ID starting with 2b79e5b58e638b0f064cac5a56ff8e687cd0db891a073c26d1444706639d567d not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.726632 4775 scope.go:117] "RemoveContainer" containerID="5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.726863 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99"} err="failed to get container status \"5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\": rpc error: code = NotFound desc = could not find container \"5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99\": container with ID starting with 5a45f98fdb325d5650858b334451985167a08d8094ca5d4764b38f065ac3be99 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.726901 4775 scope.go:117] "RemoveContainer" containerID="f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.727183 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28"} err="failed to get container status \"f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\": rpc error: code = NotFound desc = could not find container \"f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28\": container with ID starting with f010836ee58725ca90fcfe1494b3ab9f696ee327b47a27a5108136724767fa28 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.727205 4775 scope.go:117] "RemoveContainer" containerID="70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.727415 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4"} err="failed to get container status \"70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\": rpc error: code = NotFound desc = could not find container \"70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4\": container with ID starting with 70bb52f4030848dda1a76c85b6892ea6a3846de0f00e2c30c578383eacbd58f4 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.727471 4775 scope.go:117] "RemoveContainer" containerID="6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.727698 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc"} err="failed to get container status \"6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\": rpc error: code = NotFound desc = could not find container \"6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc\": container with ID starting with 6726ea0f5d8c94e4b5308e327560a3e42cc8ad012c1df59a23d9a1f9df2758cc not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.727721 4775 scope.go:117] "RemoveContainer" containerID="6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.727940 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c"} err="failed to get container status \"6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\": rpc error: code = NotFound desc = could not find container \"6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c\": container with ID starting with 6e99d8e95a7ea25573def712cc98bb5c7e2f2f69d38511bd2ab4742e91db296c not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.727961 4775 scope.go:117] "RemoveContainer" containerID="5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.728187 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde"} err="failed to get container status \"5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\": rpc error: code = NotFound desc = could not find container \"5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde\": container with ID starting with 5c1c1bdc7ba87dbea213b904e9e6422dc5a7ba2870df331886a3351a9ab6ddde not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.728224 4775 scope.go:117] "RemoveContainer" containerID="effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.728495 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6"} err="failed to get container status \"effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\": rpc error: code = NotFound desc = could not find container \"effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6\": container with ID starting with effb207e18cc86a07c45d973b406a356cac22f7c470b51edfb44907c03dddea6 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.728524 4775 scope.go:117] "RemoveContainer" containerID="294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.728857 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646"} err="failed to get container status \"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646\": rpc error: code = NotFound desc = could not find container \"294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646\": container with ID starting with 294e2a2d654e28444b7802aa1d470800d7c15c3f6dd9a43a454fb0f777f57646 not found: ID does not exist" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.755041 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bk98\" (UniqueName: \"kubernetes.io/projected/421d78dc-59bb-4b5b-9738-1eb6a6144b38-kube-api-access-5bk98\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.755083 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/421d78dc-59bb-4b5b-9738-1eb6a6144b38-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:02 crc kubenswrapper[4775]: I0321 05:00:02.755104 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/421d78dc-59bb-4b5b-9738-1eb6a6144b38-config-volume\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:03 crc kubenswrapper[4775]: I0321 05:00:03.454702 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq" Mar 21 05:00:03 crc kubenswrapper[4775]: I0321 05:00:03.454705 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq" event={"ID":"421d78dc-59bb-4b5b-9738-1eb6a6144b38","Type":"ContainerDied","Data":"639089e1b2eb45331baee2fe75955c82e4844add60351411ed01a1c3e360ac7d"} Mar 21 05:00:03 crc kubenswrapper[4775]: I0321 05:00:03.454743 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="639089e1b2eb45331baee2fe75955c82e4844add60351411ed01a1c3e360ac7d" Mar 21 05:00:03 crc kubenswrapper[4775]: I0321 05:00:03.457670 4775 generic.go:334] "Generic (PLEG): container finished" podID="82b41c52-e915-4cf3-808e-79211ba01fce" containerID="944c691144c6d64318b896ba824894bb31ce3c8bf580a8c01deaa262441a63c9" exitCode=0 Mar 21 05:00:03 crc kubenswrapper[4775]: I0321 05:00:03.457736 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" event={"ID":"82b41c52-e915-4cf3-808e-79211ba01fce","Type":"ContainerDied","Data":"944c691144c6d64318b896ba824894bb31ce3c8bf580a8c01deaa262441a63c9"} Mar 21 05:00:03 crc kubenswrapper[4775]: I0321 05:00:03.669167 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a69d31f5-deeb-4860-be96-ed5547831685" path="/var/lib/kubelet/pods/a69d31f5-deeb-4860-be96-ed5547831685/volumes" Mar 21 05:00:04 crc kubenswrapper[4775]: I0321 05:00:04.467414 4775 generic.go:334] "Generic (PLEG): container finished" podID="1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3" containerID="7d8eac46771bcd3a496da8e38bf914f7acad2a5e17c1713aea69b3a4abe8ee15" exitCode=0 Mar 21 05:00:04 crc kubenswrapper[4775]: I0321 05:00:04.467725 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567820-cngbx" event={"ID":"1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3","Type":"ContainerDied","Data":"7d8eac46771bcd3a496da8e38bf914f7acad2a5e17c1713aea69b3a4abe8ee15"} Mar 21 05:00:04 crc kubenswrapper[4775]: I0321 05:00:04.471868 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" event={"ID":"82b41c52-e915-4cf3-808e-79211ba01fce","Type":"ContainerStarted","Data":"3ce7714c061cb7e728deed8361a9b8e649c04edc7ed358977f0df05a349b7f98"} Mar 21 05:00:04 crc kubenswrapper[4775]: I0321 05:00:04.471912 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" event={"ID":"82b41c52-e915-4cf3-808e-79211ba01fce","Type":"ContainerStarted","Data":"be86b922fb3a3004ef698fa897239834e25dc954ff78dba5132dd682a8fd24c8"} Mar 21 05:00:04 crc kubenswrapper[4775]: I0321 05:00:04.471927 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" event={"ID":"82b41c52-e915-4cf3-808e-79211ba01fce","Type":"ContainerStarted","Data":"80139f9faa1db99b3a577abaf2fa3340e452f86556905d28e07464685a43c73e"} Mar 21 05:00:04 crc kubenswrapper[4775]: I0321 05:00:04.471939 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" event={"ID":"82b41c52-e915-4cf3-808e-79211ba01fce","Type":"ContainerStarted","Data":"43c4bd2d6babd3920efc2542cee5a518b18a3b87f7abfd642a1e5298d3a51c5b"} Mar 21 05:00:04 crc kubenswrapper[4775]: I0321 05:00:04.471950 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" event={"ID":"82b41c52-e915-4cf3-808e-79211ba01fce","Type":"ContainerStarted","Data":"c3fed84362c1e40dd246975c6190af593b99ae7ef29801a7f05991178af0508a"} Mar 21 05:00:04 crc kubenswrapper[4775]: I0321 05:00:04.471962 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" event={"ID":"82b41c52-e915-4cf3-808e-79211ba01fce","Type":"ContainerStarted","Data":"e0ad820fd3d42363ea591d67aeaadfee76254b32094ddf56c05606ae5ef70d74"} Mar 21 05:00:05 crc kubenswrapper[4775]: I0321 05:00:05.560606 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567820-cngbx" Mar 21 05:00:05 crc kubenswrapper[4775]: I0321 05:00:05.591313 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npphs\" (UniqueName: \"kubernetes.io/projected/1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3-kube-api-access-npphs\") pod \"1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3\" (UID: \"1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3\") " Mar 21 05:00:05 crc kubenswrapper[4775]: I0321 05:00:05.597201 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3-kube-api-access-npphs" (OuterVolumeSpecName: "kube-api-access-npphs") pod "1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3" (UID: "1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3"). InnerVolumeSpecName "kube-api-access-npphs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:00:05 crc kubenswrapper[4775]: I0321 05:00:05.693365 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npphs\" (UniqueName: \"kubernetes.io/projected/1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3-kube-api-access-npphs\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:06 crc kubenswrapper[4775]: I0321 05:00:06.490587 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567820-cngbx" event={"ID":"1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3","Type":"ContainerDied","Data":"b2b9b761df7ee4ff2e31ace5255a68fec2fe5933b83b3c71b10847ca26c3bcde"} Mar 21 05:00:06 crc kubenswrapper[4775]: I0321 05:00:06.491175 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2b9b761df7ee4ff2e31ace5255a68fec2fe5933b83b3c71b10847ca26c3bcde" Mar 21 05:00:06 crc kubenswrapper[4775]: I0321 05:00:06.490606 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567820-cngbx" Mar 21 05:00:06 crc kubenswrapper[4775]: I0321 05:00:06.495753 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" event={"ID":"82b41c52-e915-4cf3-808e-79211ba01fce","Type":"ContainerStarted","Data":"4a835e9d20b66c10c58853a57bd29b47c4c381a131c3cfe660e54e8602509a9d"} Mar 21 05:00:06 crc kubenswrapper[4775]: I0321 05:00:06.617085 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567814-fvs8k"] Mar 21 05:00:06 crc kubenswrapper[4775]: I0321 05:00:06.623595 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567814-fvs8k"] Mar 21 05:00:07 crc kubenswrapper[4775]: I0321 05:00:07.668284 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8f9e3cb-7583-4349-98dc-422a7880fe5e" path="/var/lib/kubelet/pods/e8f9e3cb-7583-4349-98dc-422a7880fe5e/volumes" Mar 21 05:00:09 crc kubenswrapper[4775]: I0321 05:00:09.514915 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" event={"ID":"82b41c52-e915-4cf3-808e-79211ba01fce","Type":"ContainerStarted","Data":"c7c7c05ece6a5336c9e4fa813e81226180e55a9c7585dc95d8cc4824c277ca42"} Mar 21 05:00:09 crc kubenswrapper[4775]: I0321 05:00:09.515612 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:09 crc kubenswrapper[4775]: I0321 05:00:09.515630 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:09 crc kubenswrapper[4775]: I0321 05:00:09.547972 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:09 crc kubenswrapper[4775]: I0321 05:00:09.553144 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" podStartSLOduration=7.5531060530000005 podStartE2EDuration="7.553106053s" podCreationTimestamp="2026-03-21 05:00:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:00:09.550836079 +0000 UTC m=+762.527299713" watchObservedRunningTime="2026-03-21 05:00:09.553106053 +0000 UTC m=+762.529569687" Mar 21 05:00:10 crc kubenswrapper[4775]: I0321 05:00:10.522105 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:10 crc kubenswrapper[4775]: I0321 05:00:10.573295 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:16 crc kubenswrapper[4775]: I0321 05:00:16.661682 4775 scope.go:117] "RemoveContainer" containerID="ad12949f26afe1756d4a6c0d01069cb26a8928bb36cc24602d4ab1bbde117f9e" Mar 21 05:00:16 crc kubenswrapper[4775]: E0321 05:00:16.662351 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-556rg_openshift-multus(e77ec218-42da-4f07-b214-184c4f3b20f3)\"" pod="openshift-multus/multus-556rg" podUID="e77ec218-42da-4f07-b214-184c4f3b20f3" Mar 21 05:00:28 crc kubenswrapper[4775]: I0321 05:00:28.661930 4775 scope.go:117] "RemoveContainer" containerID="ad12949f26afe1756d4a6c0d01069cb26a8928bb36cc24602d4ab1bbde117f9e" Mar 21 05:00:29 crc kubenswrapper[4775]: I0321 05:00:29.664841 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-556rg_e77ec218-42da-4f07-b214-184c4f3b20f3/kube-multus/2.log" Mar 21 05:00:29 crc kubenswrapper[4775]: I0321 05:00:29.666883 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-556rg_e77ec218-42da-4f07-b214-184c4f3b20f3/kube-multus/1.log" Mar 21 05:00:29 crc kubenswrapper[4775]: I0321 05:00:29.667466 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-556rg" event={"ID":"e77ec218-42da-4f07-b214-184c4f3b20f3","Type":"ContainerStarted","Data":"0db356b6d1097c58338f09b442231119fce13c2c3522c04f201fc30e7c3911fd"} Mar 21 05:00:32 crc kubenswrapper[4775]: I0321 05:00:32.421969 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-sbftc" Mar 21 05:00:32 crc kubenswrapper[4775]: I0321 05:00:32.481839 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:00:32 crc kubenswrapper[4775]: I0321 05:00:32.481910 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:00:37 crc kubenswrapper[4775]: I0321 05:00:37.179009 4775 scope.go:117] "RemoveContainer" containerID="55e31feec3221e168e396a1b14aac3afcd926bdfa911a175f057a586cb60827a" Mar 21 05:00:37 crc kubenswrapper[4775]: I0321 05:00:37.221325 4775 scope.go:117] "RemoveContainer" containerID="eaf83cd6f4a806265a871a5672f1a6939180308c2815b87ca254164f26d05aae" Mar 21 05:00:38 crc kubenswrapper[4775]: I0321 05:00:38.191040 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-556rg_e77ec218-42da-4f07-b214-184c4f3b20f3/kube-multus/2.log" Mar 21 05:00:40 crc kubenswrapper[4775]: I0321 05:00:40.123841 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp"] Mar 21 05:00:40 crc kubenswrapper[4775]: E0321 05:00:40.124459 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="421d78dc-59bb-4b5b-9738-1eb6a6144b38" containerName="collect-profiles" Mar 21 05:00:40 crc kubenswrapper[4775]: I0321 05:00:40.124476 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="421d78dc-59bb-4b5b-9738-1eb6a6144b38" containerName="collect-profiles" Mar 21 05:00:40 crc kubenswrapper[4775]: E0321 05:00:40.124500 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3" containerName="oc" Mar 21 05:00:40 crc kubenswrapper[4775]: I0321 05:00:40.124508 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3" containerName="oc" Mar 21 05:00:40 crc kubenswrapper[4775]: I0321 05:00:40.124629 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3" containerName="oc" Mar 21 05:00:40 crc kubenswrapper[4775]: I0321 05:00:40.124650 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="421d78dc-59bb-4b5b-9738-1eb6a6144b38" containerName="collect-profiles" Mar 21 05:00:40 crc kubenswrapper[4775]: I0321 05:00:40.125425 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp" Mar 21 05:00:40 crc kubenswrapper[4775]: I0321 05:00:40.128010 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 21 05:00:40 crc kubenswrapper[4775]: I0321 05:00:40.135091 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp"] Mar 21 05:00:40 crc kubenswrapper[4775]: I0321 05:00:40.219213 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rfpl\" (UniqueName: \"kubernetes.io/projected/b2ca57eb-edac-4db7-bfa4-3198d80daf99-kube-api-access-4rfpl\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp\" (UID: \"b2ca57eb-edac-4db7-bfa4-3198d80daf99\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp" Mar 21 05:00:40 crc kubenswrapper[4775]: I0321 05:00:40.219307 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b2ca57eb-edac-4db7-bfa4-3198d80daf99-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp\" (UID: \"b2ca57eb-edac-4db7-bfa4-3198d80daf99\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp" Mar 21 05:00:40 crc kubenswrapper[4775]: I0321 05:00:40.219390 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b2ca57eb-edac-4db7-bfa4-3198d80daf99-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp\" (UID: \"b2ca57eb-edac-4db7-bfa4-3198d80daf99\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp" Mar 21 05:00:40 crc kubenswrapper[4775]: I0321 05:00:40.320088 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b2ca57eb-edac-4db7-bfa4-3198d80daf99-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp\" (UID: \"b2ca57eb-edac-4db7-bfa4-3198d80daf99\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp" Mar 21 05:00:40 crc kubenswrapper[4775]: I0321 05:00:40.320203 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rfpl\" (UniqueName: \"kubernetes.io/projected/b2ca57eb-edac-4db7-bfa4-3198d80daf99-kube-api-access-4rfpl\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp\" (UID: \"b2ca57eb-edac-4db7-bfa4-3198d80daf99\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp" Mar 21 05:00:40 crc kubenswrapper[4775]: I0321 05:00:40.320225 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b2ca57eb-edac-4db7-bfa4-3198d80daf99-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp\" (UID: \"b2ca57eb-edac-4db7-bfa4-3198d80daf99\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp" Mar 21 05:00:40 crc kubenswrapper[4775]: I0321 05:00:40.320709 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b2ca57eb-edac-4db7-bfa4-3198d80daf99-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp\" (UID: \"b2ca57eb-edac-4db7-bfa4-3198d80daf99\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp" Mar 21 05:00:40 crc kubenswrapper[4775]: I0321 05:00:40.320949 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b2ca57eb-edac-4db7-bfa4-3198d80daf99-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp\" (UID: \"b2ca57eb-edac-4db7-bfa4-3198d80daf99\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp" Mar 21 05:00:40 crc kubenswrapper[4775]: I0321 05:00:40.338052 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rfpl\" (UniqueName: \"kubernetes.io/projected/b2ca57eb-edac-4db7-bfa4-3198d80daf99-kube-api-access-4rfpl\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp\" (UID: \"b2ca57eb-edac-4db7-bfa4-3198d80daf99\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp" Mar 21 05:00:40 crc kubenswrapper[4775]: I0321 05:00:40.439459 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp" Mar 21 05:00:40 crc kubenswrapper[4775]: I0321 05:00:40.620589 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp"] Mar 21 05:00:41 crc kubenswrapper[4775]: I0321 05:00:41.213096 4775 generic.go:334] "Generic (PLEG): container finished" podID="b2ca57eb-edac-4db7-bfa4-3198d80daf99" containerID="36a131f8b044d35d9bd3c10c1ce29ee06e1b450f9a0c4659063f6c4ddd7f5ec1" exitCode=0 Mar 21 05:00:41 crc kubenswrapper[4775]: I0321 05:00:41.213815 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp" event={"ID":"b2ca57eb-edac-4db7-bfa4-3198d80daf99","Type":"ContainerDied","Data":"36a131f8b044d35d9bd3c10c1ce29ee06e1b450f9a0c4659063f6c4ddd7f5ec1"} Mar 21 05:00:41 crc kubenswrapper[4775]: I0321 05:00:41.213905 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp" event={"ID":"b2ca57eb-edac-4db7-bfa4-3198d80daf99","Type":"ContainerStarted","Data":"e01461b988824f083ab26e81f6c7aec2773113cbff030f0de9786bda5a2952a6"} Mar 21 05:00:43 crc kubenswrapper[4775]: I0321 05:00:43.226165 4775 generic.go:334] "Generic (PLEG): container finished" podID="b2ca57eb-edac-4db7-bfa4-3198d80daf99" containerID="88a56d96f0d83816ad532652ef039719ef3427187d9249736e96283a01f5026a" exitCode=0 Mar 21 05:00:43 crc kubenswrapper[4775]: I0321 05:00:43.226239 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp" event={"ID":"b2ca57eb-edac-4db7-bfa4-3198d80daf99","Type":"ContainerDied","Data":"88a56d96f0d83816ad532652ef039719ef3427187d9249736e96283a01f5026a"} Mar 21 05:00:44 crc kubenswrapper[4775]: I0321 05:00:44.237899 4775 generic.go:334] "Generic (PLEG): container finished" podID="b2ca57eb-edac-4db7-bfa4-3198d80daf99" containerID="6247146119e67cb4d90c830a7ac3fc2c13acc7a731c50205e40dffabbe26a488" exitCode=0 Mar 21 05:00:44 crc kubenswrapper[4775]: I0321 05:00:44.237966 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp" event={"ID":"b2ca57eb-edac-4db7-bfa4-3198d80daf99","Type":"ContainerDied","Data":"6247146119e67cb4d90c830a7ac3fc2c13acc7a731c50205e40dffabbe26a488"} Mar 21 05:00:45 crc kubenswrapper[4775]: I0321 05:00:45.446799 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp" Mar 21 05:00:45 crc kubenswrapper[4775]: I0321 05:00:45.510967 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b2ca57eb-edac-4db7-bfa4-3198d80daf99-util\") pod \"b2ca57eb-edac-4db7-bfa4-3198d80daf99\" (UID: \"b2ca57eb-edac-4db7-bfa4-3198d80daf99\") " Mar 21 05:00:45 crc kubenswrapper[4775]: I0321 05:00:45.511035 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rfpl\" (UniqueName: \"kubernetes.io/projected/b2ca57eb-edac-4db7-bfa4-3198d80daf99-kube-api-access-4rfpl\") pod \"b2ca57eb-edac-4db7-bfa4-3198d80daf99\" (UID: \"b2ca57eb-edac-4db7-bfa4-3198d80daf99\") " Mar 21 05:00:45 crc kubenswrapper[4775]: I0321 05:00:45.511078 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b2ca57eb-edac-4db7-bfa4-3198d80daf99-bundle\") pod \"b2ca57eb-edac-4db7-bfa4-3198d80daf99\" (UID: \"b2ca57eb-edac-4db7-bfa4-3198d80daf99\") " Mar 21 05:00:45 crc kubenswrapper[4775]: I0321 05:00:45.511840 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2ca57eb-edac-4db7-bfa4-3198d80daf99-bundle" (OuterVolumeSpecName: "bundle") pod "b2ca57eb-edac-4db7-bfa4-3198d80daf99" (UID: "b2ca57eb-edac-4db7-bfa4-3198d80daf99"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:00:45 crc kubenswrapper[4775]: I0321 05:00:45.522388 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2ca57eb-edac-4db7-bfa4-3198d80daf99-kube-api-access-4rfpl" (OuterVolumeSpecName: "kube-api-access-4rfpl") pod "b2ca57eb-edac-4db7-bfa4-3198d80daf99" (UID: "b2ca57eb-edac-4db7-bfa4-3198d80daf99"). InnerVolumeSpecName "kube-api-access-4rfpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:00:45 crc kubenswrapper[4775]: I0321 05:00:45.525260 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2ca57eb-edac-4db7-bfa4-3198d80daf99-util" (OuterVolumeSpecName: "util") pod "b2ca57eb-edac-4db7-bfa4-3198d80daf99" (UID: "b2ca57eb-edac-4db7-bfa4-3198d80daf99"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:00:45 crc kubenswrapper[4775]: I0321 05:00:45.612463 4775 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b2ca57eb-edac-4db7-bfa4-3198d80daf99-util\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:45 crc kubenswrapper[4775]: I0321 05:00:45.612502 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rfpl\" (UniqueName: \"kubernetes.io/projected/b2ca57eb-edac-4db7-bfa4-3198d80daf99-kube-api-access-4rfpl\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:45 crc kubenswrapper[4775]: I0321 05:00:45.612512 4775 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b2ca57eb-edac-4db7-bfa4-3198d80daf99-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:00:46 crc kubenswrapper[4775]: I0321 05:00:46.256453 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp" event={"ID":"b2ca57eb-edac-4db7-bfa4-3198d80daf99","Type":"ContainerDied","Data":"e01461b988824f083ab26e81f6c7aec2773113cbff030f0de9786bda5a2952a6"} Mar 21 05:00:46 crc kubenswrapper[4775]: I0321 05:00:46.256546 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e01461b988824f083ab26e81f6c7aec2773113cbff030f0de9786bda5a2952a6" Mar 21 05:00:46 crc kubenswrapper[4775]: I0321 05:00:46.256564 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp" Mar 21 05:00:48 crc kubenswrapper[4775]: I0321 05:00:48.404913 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-xljlp"] Mar 21 05:00:48 crc kubenswrapper[4775]: E0321 05:00:48.405482 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ca57eb-edac-4db7-bfa4-3198d80daf99" containerName="util" Mar 21 05:00:48 crc kubenswrapper[4775]: I0321 05:00:48.405502 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ca57eb-edac-4db7-bfa4-3198d80daf99" containerName="util" Mar 21 05:00:48 crc kubenswrapper[4775]: E0321 05:00:48.405526 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ca57eb-edac-4db7-bfa4-3198d80daf99" containerName="extract" Mar 21 05:00:48 crc kubenswrapper[4775]: I0321 05:00:48.405537 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ca57eb-edac-4db7-bfa4-3198d80daf99" containerName="extract" Mar 21 05:00:48 crc kubenswrapper[4775]: E0321 05:00:48.405554 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ca57eb-edac-4db7-bfa4-3198d80daf99" containerName="pull" Mar 21 05:00:48 crc kubenswrapper[4775]: I0321 05:00:48.405564 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ca57eb-edac-4db7-bfa4-3198d80daf99" containerName="pull" Mar 21 05:00:48 crc kubenswrapper[4775]: I0321 05:00:48.405714 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2ca57eb-edac-4db7-bfa4-3198d80daf99" containerName="extract" Mar 21 05:00:48 crc kubenswrapper[4775]: I0321 05:00:48.406178 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-xljlp" Mar 21 05:00:48 crc kubenswrapper[4775]: I0321 05:00:48.407603 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-26vmx" Mar 21 05:00:48 crc kubenswrapper[4775]: I0321 05:00:48.407761 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Mar 21 05:00:48 crc kubenswrapper[4775]: I0321 05:00:48.407870 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Mar 21 05:00:48 crc kubenswrapper[4775]: I0321 05:00:48.416612 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-xljlp"] Mar 21 05:00:48 crc kubenswrapper[4775]: I0321 05:00:48.549228 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npn4x\" (UniqueName: \"kubernetes.io/projected/f4a0a79a-5b67-44f8-9ef5-304530d5e764-kube-api-access-npn4x\") pod \"nmstate-operator-796d4cfff4-xljlp\" (UID: \"f4a0a79a-5b67-44f8-9ef5-304530d5e764\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-xljlp" Mar 21 05:00:48 crc kubenswrapper[4775]: I0321 05:00:48.650656 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npn4x\" (UniqueName: \"kubernetes.io/projected/f4a0a79a-5b67-44f8-9ef5-304530d5e764-kube-api-access-npn4x\") pod \"nmstate-operator-796d4cfff4-xljlp\" (UID: \"f4a0a79a-5b67-44f8-9ef5-304530d5e764\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-xljlp" Mar 21 05:00:48 crc kubenswrapper[4775]: I0321 05:00:48.669908 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npn4x\" (UniqueName: \"kubernetes.io/projected/f4a0a79a-5b67-44f8-9ef5-304530d5e764-kube-api-access-npn4x\") pod \"nmstate-operator-796d4cfff4-xljlp\" (UID: \"f4a0a79a-5b67-44f8-9ef5-304530d5e764\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-xljlp" Mar 21 05:00:48 crc kubenswrapper[4775]: I0321 05:00:48.721554 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-xljlp" Mar 21 05:00:48 crc kubenswrapper[4775]: I0321 05:00:48.915437 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-xljlp"] Mar 21 05:00:49 crc kubenswrapper[4775]: I0321 05:00:49.273248 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-xljlp" event={"ID":"f4a0a79a-5b67-44f8-9ef5-304530d5e764","Type":"ContainerStarted","Data":"a85340a544736d29c8ba4919cf40b85e8a8702c0d41080670e1974a9f1db9478"} Mar 21 05:00:52 crc kubenswrapper[4775]: I0321 05:00:52.299068 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-xljlp" event={"ID":"f4a0a79a-5b67-44f8-9ef5-304530d5e764","Type":"ContainerStarted","Data":"d17303fb14c393392fb271c5ccb948311cc98edae2519ee1050ffcbb981fdfb3"} Mar 21 05:00:52 crc kubenswrapper[4775]: I0321 05:00:52.321251 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-796d4cfff4-xljlp" podStartSLOduration=1.691041025 podStartE2EDuration="4.32123131s" podCreationTimestamp="2026-03-21 05:00:48 +0000 UTC" firstStartedPulling="2026-03-21 05:00:48.919628472 +0000 UTC m=+801.896092096" lastFinishedPulling="2026-03-21 05:00:51.549818757 +0000 UTC m=+804.526282381" observedRunningTime="2026-03-21 05:00:52.318066301 +0000 UTC m=+805.294529935" watchObservedRunningTime="2026-03-21 05:00:52.32123131 +0000 UTC m=+805.297694934" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.195440 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-626kh"] Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.196832 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-626kh" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.198972 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-nkc28" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.211272 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-626kh"] Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.216084 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-rptn7"] Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.216889 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-rptn7" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.219885 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.229150 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-rptn7"] Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.243935 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-6l74f"] Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.244759 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-6l74f" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.309886 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d03e5939-1625-4597-ad3b-9edf8e8075f5-ovs-socket\") pod \"nmstate-handler-6l74f\" (UID: \"d03e5939-1625-4597-ad3b-9edf8e8075f5\") " pod="openshift-nmstate/nmstate-handler-6l74f" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.309928 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d03e5939-1625-4597-ad3b-9edf8e8075f5-dbus-socket\") pod \"nmstate-handler-6l74f\" (UID: \"d03e5939-1625-4597-ad3b-9edf8e8075f5\") " pod="openshift-nmstate/nmstate-handler-6l74f" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.309952 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d03e5939-1625-4597-ad3b-9edf8e8075f5-nmstate-lock\") pod \"nmstate-handler-6l74f\" (UID: \"d03e5939-1625-4597-ad3b-9edf8e8075f5\") " pod="openshift-nmstate/nmstate-handler-6l74f" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.309974 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwxxf\" (UniqueName: \"kubernetes.io/projected/d03e5939-1625-4597-ad3b-9edf8e8075f5-kube-api-access-bwxxf\") pod \"nmstate-handler-6l74f\" (UID: \"d03e5939-1625-4597-ad3b-9edf8e8075f5\") " pod="openshift-nmstate/nmstate-handler-6l74f" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.310005 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drl6w\" (UniqueName: \"kubernetes.io/projected/409422f2-717f-4f82-8dae-fe01dfda7083-kube-api-access-drl6w\") pod \"nmstate-metrics-9b8c8685d-626kh\" (UID: \"409422f2-717f-4f82-8dae-fe01dfda7083\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-626kh" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.310043 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/944b76e5-c8c5-4cba-9df2-9e9b87b540a8-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-rptn7\" (UID: \"944b76e5-c8c5-4cba-9df2-9e9b87b540a8\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-rptn7" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.310063 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfmp7\" (UniqueName: \"kubernetes.io/projected/944b76e5-c8c5-4cba-9df2-9e9b87b540a8-kube-api-access-lfmp7\") pod \"nmstate-webhook-5f558f5558-rptn7\" (UID: \"944b76e5-c8c5-4cba-9df2-9e9b87b540a8\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-rptn7" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.344182 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-8x7x2"] Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.344890 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-8x7x2" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.347780 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-nsznh" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.348019 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.348176 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.355698 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-8x7x2"] Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.411314 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0fbab95-1c88-40a3-8ccb-58bca74c8f3c-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-8x7x2\" (UID: \"b0fbab95-1c88-40a3-8ccb-58bca74c8f3c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-8x7x2" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.411394 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d97vw\" (UniqueName: \"kubernetes.io/projected/b0fbab95-1c88-40a3-8ccb-58bca74c8f3c-kube-api-access-d97vw\") pod \"nmstate-console-plugin-86f58fcf4-8x7x2\" (UID: \"b0fbab95-1c88-40a3-8ccb-58bca74c8f3c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-8x7x2" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.411454 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d03e5939-1625-4597-ad3b-9edf8e8075f5-ovs-socket\") pod \"nmstate-handler-6l74f\" (UID: \"d03e5939-1625-4597-ad3b-9edf8e8075f5\") " pod="openshift-nmstate/nmstate-handler-6l74f" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.411531 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d03e5939-1625-4597-ad3b-9edf8e8075f5-dbus-socket\") pod \"nmstate-handler-6l74f\" (UID: \"d03e5939-1625-4597-ad3b-9edf8e8075f5\") " pod="openshift-nmstate/nmstate-handler-6l74f" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.411609 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d03e5939-1625-4597-ad3b-9edf8e8075f5-nmstate-lock\") pod \"nmstate-handler-6l74f\" (UID: \"d03e5939-1625-4597-ad3b-9edf8e8075f5\") " pod="openshift-nmstate/nmstate-handler-6l74f" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.411643 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwxxf\" (UniqueName: \"kubernetes.io/projected/d03e5939-1625-4597-ad3b-9edf8e8075f5-kube-api-access-bwxxf\") pod \"nmstate-handler-6l74f\" (UID: \"d03e5939-1625-4597-ad3b-9edf8e8075f5\") " pod="openshift-nmstate/nmstate-handler-6l74f" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.411668 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b0fbab95-1c88-40a3-8ccb-58bca74c8f3c-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-8x7x2\" (UID: \"b0fbab95-1c88-40a3-8ccb-58bca74c8f3c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-8x7x2" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.411698 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drl6w\" (UniqueName: \"kubernetes.io/projected/409422f2-717f-4f82-8dae-fe01dfda7083-kube-api-access-drl6w\") pod \"nmstate-metrics-9b8c8685d-626kh\" (UID: \"409422f2-717f-4f82-8dae-fe01dfda7083\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-626kh" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.411739 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfmp7\" (UniqueName: \"kubernetes.io/projected/944b76e5-c8c5-4cba-9df2-9e9b87b540a8-kube-api-access-lfmp7\") pod \"nmstate-webhook-5f558f5558-rptn7\" (UID: \"944b76e5-c8c5-4cba-9df2-9e9b87b540a8\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-rptn7" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.411757 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/944b76e5-c8c5-4cba-9df2-9e9b87b540a8-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-rptn7\" (UID: \"944b76e5-c8c5-4cba-9df2-9e9b87b540a8\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-rptn7" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.412084 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d03e5939-1625-4597-ad3b-9edf8e8075f5-nmstate-lock\") pod \"nmstate-handler-6l74f\" (UID: \"d03e5939-1625-4597-ad3b-9edf8e8075f5\") " pod="openshift-nmstate/nmstate-handler-6l74f" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.412143 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d03e5939-1625-4597-ad3b-9edf8e8075f5-ovs-socket\") pod \"nmstate-handler-6l74f\" (UID: \"d03e5939-1625-4597-ad3b-9edf8e8075f5\") " pod="openshift-nmstate/nmstate-handler-6l74f" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.412201 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d03e5939-1625-4597-ad3b-9edf8e8075f5-dbus-socket\") pod \"nmstate-handler-6l74f\" (UID: \"d03e5939-1625-4597-ad3b-9edf8e8075f5\") " pod="openshift-nmstate/nmstate-handler-6l74f" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.420793 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/944b76e5-c8c5-4cba-9df2-9e9b87b540a8-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-rptn7\" (UID: \"944b76e5-c8c5-4cba-9df2-9e9b87b540a8\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-rptn7" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.430664 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drl6w\" (UniqueName: \"kubernetes.io/projected/409422f2-717f-4f82-8dae-fe01dfda7083-kube-api-access-drl6w\") pod \"nmstate-metrics-9b8c8685d-626kh\" (UID: \"409422f2-717f-4f82-8dae-fe01dfda7083\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-626kh" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.433356 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfmp7\" (UniqueName: \"kubernetes.io/projected/944b76e5-c8c5-4cba-9df2-9e9b87b540a8-kube-api-access-lfmp7\") pod \"nmstate-webhook-5f558f5558-rptn7\" (UID: \"944b76e5-c8c5-4cba-9df2-9e9b87b540a8\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-rptn7" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.435391 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwxxf\" (UniqueName: \"kubernetes.io/projected/d03e5939-1625-4597-ad3b-9edf8e8075f5-kube-api-access-bwxxf\") pod \"nmstate-handler-6l74f\" (UID: \"d03e5939-1625-4597-ad3b-9edf8e8075f5\") " pod="openshift-nmstate/nmstate-handler-6l74f" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.513214 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b0fbab95-1c88-40a3-8ccb-58bca74c8f3c-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-8x7x2\" (UID: \"b0fbab95-1c88-40a3-8ccb-58bca74c8f3c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-8x7x2" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.513297 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0fbab95-1c88-40a3-8ccb-58bca74c8f3c-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-8x7x2\" (UID: \"b0fbab95-1c88-40a3-8ccb-58bca74c8f3c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-8x7x2" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.513351 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d97vw\" (UniqueName: \"kubernetes.io/projected/b0fbab95-1c88-40a3-8ccb-58bca74c8f3c-kube-api-access-d97vw\") pod \"nmstate-console-plugin-86f58fcf4-8x7x2\" (UID: \"b0fbab95-1c88-40a3-8ccb-58bca74c8f3c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-8x7x2" Mar 21 05:00:53 crc kubenswrapper[4775]: E0321 05:00:53.513732 4775 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Mar 21 05:00:53 crc kubenswrapper[4775]: E0321 05:00:53.513808 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b0fbab95-1c88-40a3-8ccb-58bca74c8f3c-plugin-serving-cert podName:b0fbab95-1c88-40a3-8ccb-58bca74c8f3c nodeName:}" failed. No retries permitted until 2026-03-21 05:00:54.013788572 +0000 UTC m=+806.990252196 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/b0fbab95-1c88-40a3-8ccb-58bca74c8f3c-plugin-serving-cert") pod "nmstate-console-plugin-86f58fcf4-8x7x2" (UID: "b0fbab95-1c88-40a3-8ccb-58bca74c8f3c") : secret "plugin-serving-cert" not found Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.514077 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b0fbab95-1c88-40a3-8ccb-58bca74c8f3c-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-8x7x2\" (UID: \"b0fbab95-1c88-40a3-8ccb-58bca74c8f3c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-8x7x2" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.522725 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-626kh" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.530415 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6f5cb9cdf8-7rs8l"] Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.531353 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.535688 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-rptn7" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.535980 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d97vw\" (UniqueName: \"kubernetes.io/projected/b0fbab95-1c88-40a3-8ccb-58bca74c8f3c-kube-api-access-d97vw\") pod \"nmstate-console-plugin-86f58fcf4-8x7x2\" (UID: \"b0fbab95-1c88-40a3-8ccb-58bca74c8f3c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-8x7x2" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.554608 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6f5cb9cdf8-7rs8l"] Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.567990 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-6l74f" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.613903 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/552a8ec9-0977-4124-8cec-00e0b85fbc8d-service-ca\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.613943 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/552a8ec9-0977-4124-8cec-00e0b85fbc8d-console-oauth-config\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.613971 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/552a8ec9-0977-4124-8cec-00e0b85fbc8d-oauth-serving-cert\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.621552 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/552a8ec9-0977-4124-8cec-00e0b85fbc8d-trusted-ca-bundle\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.621675 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx6ns\" (UniqueName: \"kubernetes.io/projected/552a8ec9-0977-4124-8cec-00e0b85fbc8d-kube-api-access-lx6ns\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.621755 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/552a8ec9-0977-4124-8cec-00e0b85fbc8d-console-serving-cert\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.621786 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/552a8ec9-0977-4124-8cec-00e0b85fbc8d-console-config\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.723489 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/552a8ec9-0977-4124-8cec-00e0b85fbc8d-service-ca\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.723571 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/552a8ec9-0977-4124-8cec-00e0b85fbc8d-console-oauth-config\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.723728 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/552a8ec9-0977-4124-8cec-00e0b85fbc8d-oauth-serving-cert\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.723794 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/552a8ec9-0977-4124-8cec-00e0b85fbc8d-trusted-ca-bundle\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.723851 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx6ns\" (UniqueName: \"kubernetes.io/projected/552a8ec9-0977-4124-8cec-00e0b85fbc8d-kube-api-access-lx6ns\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.723904 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/552a8ec9-0977-4124-8cec-00e0b85fbc8d-console-serving-cert\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.723926 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/552a8ec9-0977-4124-8cec-00e0b85fbc8d-console-config\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.724587 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/552a8ec9-0977-4124-8cec-00e0b85fbc8d-service-ca\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.724615 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/552a8ec9-0977-4124-8cec-00e0b85fbc8d-oauth-serving-cert\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.724977 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/552a8ec9-0977-4124-8cec-00e0b85fbc8d-console-config\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.725098 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/552a8ec9-0977-4124-8cec-00e0b85fbc8d-trusted-ca-bundle\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.729676 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/552a8ec9-0977-4124-8cec-00e0b85fbc8d-console-oauth-config\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.729804 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/552a8ec9-0977-4124-8cec-00e0b85fbc8d-console-serving-cert\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.739599 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx6ns\" (UniqueName: \"kubernetes.io/projected/552a8ec9-0977-4124-8cec-00e0b85fbc8d-kube-api-access-lx6ns\") pod \"console-6f5cb9cdf8-7rs8l\" (UID: \"552a8ec9-0977-4124-8cec-00e0b85fbc8d\") " pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.768151 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-626kh"] Mar 21 05:00:53 crc kubenswrapper[4775]: W0321 05:00:53.768309 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod409422f2_717f_4f82_8dae_fe01dfda7083.slice/crio-1d3267450edd9496263daf21110ec6f9a451661a7cefefc5ae4e58832a5cb02b WatchSource:0}: Error finding container 1d3267450edd9496263daf21110ec6f9a451661a7cefefc5ae4e58832a5cb02b: Status 404 returned error can't find the container with id 1d3267450edd9496263daf21110ec6f9a451661a7cefefc5ae4e58832a5cb02b Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.809794 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-rptn7"] Mar 21 05:00:53 crc kubenswrapper[4775]: W0321 05:00:53.810173 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod944b76e5_c8c5_4cba_9df2_9e9b87b540a8.slice/crio-90555b21c6e51115e21138f33eb614f4dd3c6518bac50d302dd41c2851a3570e WatchSource:0}: Error finding container 90555b21c6e51115e21138f33eb614f4dd3c6518bac50d302dd41c2851a3570e: Status 404 returned error can't find the container with id 90555b21c6e51115e21138f33eb614f4dd3c6518bac50d302dd41c2851a3570e Mar 21 05:00:53 crc kubenswrapper[4775]: I0321 05:00:53.908173 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:00:54 crc kubenswrapper[4775]: I0321 05:00:54.031244 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0fbab95-1c88-40a3-8ccb-58bca74c8f3c-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-8x7x2\" (UID: \"b0fbab95-1c88-40a3-8ccb-58bca74c8f3c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-8x7x2" Mar 21 05:00:54 crc kubenswrapper[4775]: I0321 05:00:54.036017 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0fbab95-1c88-40a3-8ccb-58bca74c8f3c-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-8x7x2\" (UID: \"b0fbab95-1c88-40a3-8ccb-58bca74c8f3c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-8x7x2" Mar 21 05:00:54 crc kubenswrapper[4775]: I0321 05:00:54.110752 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6f5cb9cdf8-7rs8l"] Mar 21 05:00:54 crc kubenswrapper[4775]: W0321 05:00:54.116239 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod552a8ec9_0977_4124_8cec_00e0b85fbc8d.slice/crio-fc9d44e29491eb027572cbe53cb1eff5ccf0b664f29695d8334d930979423a33 WatchSource:0}: Error finding container fc9d44e29491eb027572cbe53cb1eff5ccf0b664f29695d8334d930979423a33: Status 404 returned error can't find the container with id fc9d44e29491eb027572cbe53cb1eff5ccf0b664f29695d8334d930979423a33 Mar 21 05:00:54 crc kubenswrapper[4775]: I0321 05:00:54.260750 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-8x7x2" Mar 21 05:00:54 crc kubenswrapper[4775]: I0321 05:00:54.317886 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-626kh" event={"ID":"409422f2-717f-4f82-8dae-fe01dfda7083","Type":"ContainerStarted","Data":"1d3267450edd9496263daf21110ec6f9a451661a7cefefc5ae4e58832a5cb02b"} Mar 21 05:00:54 crc kubenswrapper[4775]: I0321 05:00:54.320516 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6f5cb9cdf8-7rs8l" event={"ID":"552a8ec9-0977-4124-8cec-00e0b85fbc8d","Type":"ContainerStarted","Data":"3e7a82b84b89b9c7236ebc322b6f2a7fec36ac042a44ffdc020c14e50603242d"} Mar 21 05:00:54 crc kubenswrapper[4775]: I0321 05:00:54.320736 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6f5cb9cdf8-7rs8l" event={"ID":"552a8ec9-0977-4124-8cec-00e0b85fbc8d","Type":"ContainerStarted","Data":"fc9d44e29491eb027572cbe53cb1eff5ccf0b664f29695d8334d930979423a33"} Mar 21 05:00:54 crc kubenswrapper[4775]: I0321 05:00:54.322889 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-6l74f" event={"ID":"d03e5939-1625-4597-ad3b-9edf8e8075f5","Type":"ContainerStarted","Data":"15538dc9766f3a74b3af18a8ff5b6cbca13dbed2cdd76c6e407ab128491aef08"} Mar 21 05:00:54 crc kubenswrapper[4775]: I0321 05:00:54.326289 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-rptn7" event={"ID":"944b76e5-c8c5-4cba-9df2-9e9b87b540a8","Type":"ContainerStarted","Data":"90555b21c6e51115e21138f33eb614f4dd3c6518bac50d302dd41c2851a3570e"} Mar 21 05:00:54 crc kubenswrapper[4775]: I0321 05:00:54.341388 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6f5cb9cdf8-7rs8l" podStartSLOduration=1.341368592 podStartE2EDuration="1.341368592s" podCreationTimestamp="2026-03-21 05:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:00:54.340694023 +0000 UTC m=+807.317157647" watchObservedRunningTime="2026-03-21 05:00:54.341368592 +0000 UTC m=+807.317832206" Mar 21 05:00:54 crc kubenswrapper[4775]: I0321 05:00:54.538597 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-8x7x2"] Mar 21 05:00:54 crc kubenswrapper[4775]: W0321 05:00:54.546715 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0fbab95_1c88_40a3_8ccb_58bca74c8f3c.slice/crio-a023ff7a200326a4a00013d768ba6028b20e32eb59ada071ba4db2fdde9bf6fb WatchSource:0}: Error finding container a023ff7a200326a4a00013d768ba6028b20e32eb59ada071ba4db2fdde9bf6fb: Status 404 returned error can't find the container with id a023ff7a200326a4a00013d768ba6028b20e32eb59ada071ba4db2fdde9bf6fb Mar 21 05:00:55 crc kubenswrapper[4775]: I0321 05:00:55.332597 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-8x7x2" event={"ID":"b0fbab95-1c88-40a3-8ccb-58bca74c8f3c","Type":"ContainerStarted","Data":"a023ff7a200326a4a00013d768ba6028b20e32eb59ada071ba4db2fdde9bf6fb"} Mar 21 05:00:57 crc kubenswrapper[4775]: I0321 05:00:57.349248 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-6l74f" event={"ID":"d03e5939-1625-4597-ad3b-9edf8e8075f5","Type":"ContainerStarted","Data":"c8f8b3621d8aebe8542eaca9c486a0394dea6add6674b43c81c01465cd4477be"} Mar 21 05:00:57 crc kubenswrapper[4775]: I0321 05:00:57.350258 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-6l74f" Mar 21 05:00:57 crc kubenswrapper[4775]: I0321 05:00:57.351771 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-rptn7" event={"ID":"944b76e5-c8c5-4cba-9df2-9e9b87b540a8","Type":"ContainerStarted","Data":"0b510fbd54cb404ec5225713a1e467cb92b4867514763b60e365ae51a7844e1d"} Mar 21 05:00:57 crc kubenswrapper[4775]: I0321 05:00:57.352398 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f558f5558-rptn7" Mar 21 05:00:57 crc kubenswrapper[4775]: I0321 05:00:57.355045 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-626kh" event={"ID":"409422f2-717f-4f82-8dae-fe01dfda7083","Type":"ContainerStarted","Data":"5db6f5f91bd6657bafe597321f3623e2d7d3343cf1edbf44c8c79b9357dfb233"} Mar 21 05:00:57 crc kubenswrapper[4775]: I0321 05:00:57.369990 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-6l74f" podStartSLOduration=1.642006135 podStartE2EDuration="4.369961092s" podCreationTimestamp="2026-03-21 05:00:53 +0000 UTC" firstStartedPulling="2026-03-21 05:00:53.622208035 +0000 UTC m=+806.598671659" lastFinishedPulling="2026-03-21 05:00:56.350162952 +0000 UTC m=+809.326626616" observedRunningTime="2026-03-21 05:00:57.365915768 +0000 UTC m=+810.342379422" watchObservedRunningTime="2026-03-21 05:00:57.369961092 +0000 UTC m=+810.346424716" Mar 21 05:00:57 crc kubenswrapper[4775]: I0321 05:00:57.390741 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f558f5558-rptn7" podStartSLOduration=1.8983759980000001 podStartE2EDuration="4.390723199s" podCreationTimestamp="2026-03-21 05:00:53 +0000 UTC" firstStartedPulling="2026-03-21 05:00:53.811724989 +0000 UTC m=+806.788188613" lastFinishedPulling="2026-03-21 05:00:56.30407219 +0000 UTC m=+809.280535814" observedRunningTime="2026-03-21 05:00:57.388792335 +0000 UTC m=+810.365255959" watchObservedRunningTime="2026-03-21 05:00:57.390723199 +0000 UTC m=+810.367186843" Mar 21 05:00:58 crc kubenswrapper[4775]: I0321 05:00:58.364112 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-8x7x2" event={"ID":"b0fbab95-1c88-40a3-8ccb-58bca74c8f3c","Type":"ContainerStarted","Data":"479c50c0568fe490b9c0c88db80fe55cd43a3d41946da6328955d3ae544b6912"} Mar 21 05:00:58 crc kubenswrapper[4775]: I0321 05:00:58.390014 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-8x7x2" podStartSLOduration=2.665570981 podStartE2EDuration="5.389995839s" podCreationTimestamp="2026-03-21 05:00:53 +0000 UTC" firstStartedPulling="2026-03-21 05:00:54.548624697 +0000 UTC m=+807.525088321" lastFinishedPulling="2026-03-21 05:00:57.273049555 +0000 UTC m=+810.249513179" observedRunningTime="2026-03-21 05:00:58.378430872 +0000 UTC m=+811.354894516" watchObservedRunningTime="2026-03-21 05:00:58.389995839 +0000 UTC m=+811.366459463" Mar 21 05:00:59 crc kubenswrapper[4775]: I0321 05:00:59.372078 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-626kh" event={"ID":"409422f2-717f-4f82-8dae-fe01dfda7083","Type":"ContainerStarted","Data":"513ac8f75e5d22ae65ecdd4bda0fd91706a6f342c292c674620878f6890dc19d"} Mar 21 05:00:59 crc kubenswrapper[4775]: I0321 05:00:59.401743 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-626kh" podStartSLOduration=1.586026514 podStartE2EDuration="6.401720541s" podCreationTimestamp="2026-03-21 05:00:53 +0000 UTC" firstStartedPulling="2026-03-21 05:00:53.77070197 +0000 UTC m=+806.747165604" lastFinishedPulling="2026-03-21 05:00:58.586396017 +0000 UTC m=+811.562859631" observedRunningTime="2026-03-21 05:00:59.399162239 +0000 UTC m=+812.375625883" watchObservedRunningTime="2026-03-21 05:00:59.401720541 +0000 UTC m=+812.378184185" Mar 21 05:01:02 crc kubenswrapper[4775]: I0321 05:01:02.482378 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:01:02 crc kubenswrapper[4775]: I0321 05:01:02.483040 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:01:03 crc kubenswrapper[4775]: I0321 05:01:03.599817 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-6l74f" Mar 21 05:01:03 crc kubenswrapper[4775]: I0321 05:01:03.908825 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:01:03 crc kubenswrapper[4775]: I0321 05:01:03.908933 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:01:03 crc kubenswrapper[4775]: I0321 05:01:03.916248 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:01:04 crc kubenswrapper[4775]: I0321 05:01:04.424081 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6f5cb9cdf8-7rs8l" Mar 21 05:01:04 crc kubenswrapper[4775]: I0321 05:01:04.523797 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-wst2s"] Mar 21 05:01:13 crc kubenswrapper[4775]: I0321 05:01:13.542634 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f558f5558-rptn7" Mar 21 05:01:20 crc kubenswrapper[4775]: I0321 05:01:20.718827 4775 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 21 05:01:25 crc kubenswrapper[4775]: I0321 05:01:25.385111 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s"] Mar 21 05:01:25 crc kubenswrapper[4775]: I0321 05:01:25.387827 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s" Mar 21 05:01:25 crc kubenswrapper[4775]: I0321 05:01:25.389356 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 21 05:01:25 crc kubenswrapper[4775]: I0321 05:01:25.396525 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s"] Mar 21 05:01:25 crc kubenswrapper[4775]: I0321 05:01:25.515595 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6252c43a-d149-46f7-ac6c-263c34980fe2-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s\" (UID: \"6252c43a-d149-46f7-ac6c-263c34980fe2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s" Mar 21 05:01:25 crc kubenswrapper[4775]: I0321 05:01:25.515661 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqtp7\" (UniqueName: \"kubernetes.io/projected/6252c43a-d149-46f7-ac6c-263c34980fe2-kube-api-access-cqtp7\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s\" (UID: \"6252c43a-d149-46f7-ac6c-263c34980fe2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s" Mar 21 05:01:25 crc kubenswrapper[4775]: I0321 05:01:25.515692 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6252c43a-d149-46f7-ac6c-263c34980fe2-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s\" (UID: \"6252c43a-d149-46f7-ac6c-263c34980fe2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s" Mar 21 05:01:25 crc kubenswrapper[4775]: I0321 05:01:25.617475 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6252c43a-d149-46f7-ac6c-263c34980fe2-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s\" (UID: \"6252c43a-d149-46f7-ac6c-263c34980fe2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s" Mar 21 05:01:25 crc kubenswrapper[4775]: I0321 05:01:25.617890 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqtp7\" (UniqueName: \"kubernetes.io/projected/6252c43a-d149-46f7-ac6c-263c34980fe2-kube-api-access-cqtp7\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s\" (UID: \"6252c43a-d149-46f7-ac6c-263c34980fe2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s" Mar 21 05:01:25 crc kubenswrapper[4775]: I0321 05:01:25.617938 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6252c43a-d149-46f7-ac6c-263c34980fe2-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s\" (UID: \"6252c43a-d149-46f7-ac6c-263c34980fe2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s" Mar 21 05:01:25 crc kubenswrapper[4775]: I0321 05:01:25.618744 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6252c43a-d149-46f7-ac6c-263c34980fe2-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s\" (UID: \"6252c43a-d149-46f7-ac6c-263c34980fe2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s" Mar 21 05:01:25 crc kubenswrapper[4775]: I0321 05:01:25.619033 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6252c43a-d149-46f7-ac6c-263c34980fe2-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s\" (UID: \"6252c43a-d149-46f7-ac6c-263c34980fe2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s" Mar 21 05:01:25 crc kubenswrapper[4775]: I0321 05:01:25.648006 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqtp7\" (UniqueName: \"kubernetes.io/projected/6252c43a-d149-46f7-ac6c-263c34980fe2-kube-api-access-cqtp7\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s\" (UID: \"6252c43a-d149-46f7-ac6c-263c34980fe2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s" Mar 21 05:01:25 crc kubenswrapper[4775]: I0321 05:01:25.702959 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s" Mar 21 05:01:25 crc kubenswrapper[4775]: I0321 05:01:25.925720 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s"] Mar 21 05:01:26 crc kubenswrapper[4775]: I0321 05:01:26.579426 4775 generic.go:334] "Generic (PLEG): container finished" podID="6252c43a-d149-46f7-ac6c-263c34980fe2" containerID="d003d37849ef559a0570ea5750927b9875f1a10ae825ecea7f42bb3392bd2ce8" exitCode=0 Mar 21 05:01:26 crc kubenswrapper[4775]: I0321 05:01:26.579503 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s" event={"ID":"6252c43a-d149-46f7-ac6c-263c34980fe2","Type":"ContainerDied","Data":"d003d37849ef559a0570ea5750927b9875f1a10ae825ecea7f42bb3392bd2ce8"} Mar 21 05:01:26 crc kubenswrapper[4775]: I0321 05:01:26.579544 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s" event={"ID":"6252c43a-d149-46f7-ac6c-263c34980fe2","Type":"ContainerStarted","Data":"9848d056d20a28e5d15c2a52eb1b9b4a66212a967ad7a7690bd7227caa27e415"} Mar 21 05:01:27 crc kubenswrapper[4775]: I0321 05:01:27.505914 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w5l6s"] Mar 21 05:01:27 crc kubenswrapper[4775]: I0321 05:01:27.507780 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w5l6s" Mar 21 05:01:27 crc kubenswrapper[4775]: I0321 05:01:27.521240 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w5l6s"] Mar 21 05:01:27 crc kubenswrapper[4775]: I0321 05:01:27.644457 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ww7q\" (UniqueName: \"kubernetes.io/projected/e3a4716c-b33b-4030-a4fa-0feda18a8657-kube-api-access-5ww7q\") pod \"redhat-operators-w5l6s\" (UID: \"e3a4716c-b33b-4030-a4fa-0feda18a8657\") " pod="openshift-marketplace/redhat-operators-w5l6s" Mar 21 05:01:27 crc kubenswrapper[4775]: I0321 05:01:27.644537 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3a4716c-b33b-4030-a4fa-0feda18a8657-utilities\") pod \"redhat-operators-w5l6s\" (UID: \"e3a4716c-b33b-4030-a4fa-0feda18a8657\") " pod="openshift-marketplace/redhat-operators-w5l6s" Mar 21 05:01:27 crc kubenswrapper[4775]: I0321 05:01:27.644626 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3a4716c-b33b-4030-a4fa-0feda18a8657-catalog-content\") pod \"redhat-operators-w5l6s\" (UID: \"e3a4716c-b33b-4030-a4fa-0feda18a8657\") " pod="openshift-marketplace/redhat-operators-w5l6s" Mar 21 05:01:27 crc kubenswrapper[4775]: I0321 05:01:27.746257 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3a4716c-b33b-4030-a4fa-0feda18a8657-catalog-content\") pod \"redhat-operators-w5l6s\" (UID: \"e3a4716c-b33b-4030-a4fa-0feda18a8657\") " pod="openshift-marketplace/redhat-operators-w5l6s" Mar 21 05:01:27 crc kubenswrapper[4775]: I0321 05:01:27.746315 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ww7q\" (UniqueName: \"kubernetes.io/projected/e3a4716c-b33b-4030-a4fa-0feda18a8657-kube-api-access-5ww7q\") pod \"redhat-operators-w5l6s\" (UID: \"e3a4716c-b33b-4030-a4fa-0feda18a8657\") " pod="openshift-marketplace/redhat-operators-w5l6s" Mar 21 05:01:27 crc kubenswrapper[4775]: I0321 05:01:27.746389 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3a4716c-b33b-4030-a4fa-0feda18a8657-utilities\") pod \"redhat-operators-w5l6s\" (UID: \"e3a4716c-b33b-4030-a4fa-0feda18a8657\") " pod="openshift-marketplace/redhat-operators-w5l6s" Mar 21 05:01:27 crc kubenswrapper[4775]: I0321 05:01:27.746881 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3a4716c-b33b-4030-a4fa-0feda18a8657-catalog-content\") pod \"redhat-operators-w5l6s\" (UID: \"e3a4716c-b33b-4030-a4fa-0feda18a8657\") " pod="openshift-marketplace/redhat-operators-w5l6s" Mar 21 05:01:27 crc kubenswrapper[4775]: I0321 05:01:27.746997 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3a4716c-b33b-4030-a4fa-0feda18a8657-utilities\") pod \"redhat-operators-w5l6s\" (UID: \"e3a4716c-b33b-4030-a4fa-0feda18a8657\") " pod="openshift-marketplace/redhat-operators-w5l6s" Mar 21 05:01:27 crc kubenswrapper[4775]: I0321 05:01:27.770091 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ww7q\" (UniqueName: \"kubernetes.io/projected/e3a4716c-b33b-4030-a4fa-0feda18a8657-kube-api-access-5ww7q\") pod \"redhat-operators-w5l6s\" (UID: \"e3a4716c-b33b-4030-a4fa-0feda18a8657\") " pod="openshift-marketplace/redhat-operators-w5l6s" Mar 21 05:01:27 crc kubenswrapper[4775]: I0321 05:01:27.827397 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w5l6s" Mar 21 05:01:28 crc kubenswrapper[4775]: I0321 05:01:28.056530 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w5l6s"] Mar 21 05:01:28 crc kubenswrapper[4775]: I0321 05:01:28.593666 4775 generic.go:334] "Generic (PLEG): container finished" podID="e3a4716c-b33b-4030-a4fa-0feda18a8657" containerID="2adc93f631ab51a3bffbfd6d7c28133602106d191f43a08a57386749aee6506a" exitCode=0 Mar 21 05:01:28 crc kubenswrapper[4775]: I0321 05:01:28.593718 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5l6s" event={"ID":"e3a4716c-b33b-4030-a4fa-0feda18a8657","Type":"ContainerDied","Data":"2adc93f631ab51a3bffbfd6d7c28133602106d191f43a08a57386749aee6506a"} Mar 21 05:01:28 crc kubenswrapper[4775]: I0321 05:01:28.593936 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5l6s" event={"ID":"e3a4716c-b33b-4030-a4fa-0feda18a8657","Type":"ContainerStarted","Data":"e0c863098fc3ee0671df4688c9648d37c09a40f58d77f06335e48b368ac9c588"} Mar 21 05:01:29 crc kubenswrapper[4775]: I0321 05:01:29.561630 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-wst2s" podUID="a9ed1e0e-eff9-4690-bcf5-45f6074c200e" containerName="console" containerID="cri-o://4843fbb3995c43ae7854ec367e570829c36ae201e54901d87c5c625dcdce3764" gracePeriod=15 Mar 21 05:01:29 crc kubenswrapper[4775]: I0321 05:01:29.602624 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5l6s" event={"ID":"e3a4716c-b33b-4030-a4fa-0feda18a8657","Type":"ContainerStarted","Data":"bfefe0cac35a2f3c7e6293ed113af6198ac9e4db485c73a73295e5e65ba7c34b"} Mar 21 05:01:29 crc kubenswrapper[4775]: I0321 05:01:29.605813 4775 generic.go:334] "Generic (PLEG): container finished" podID="6252c43a-d149-46f7-ac6c-263c34980fe2" containerID="2af57a8d89d834477d936f1accea68294ed6669b9cde85ec3a277a286c70ebc0" exitCode=0 Mar 21 05:01:29 crc kubenswrapper[4775]: I0321 05:01:29.605847 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s" event={"ID":"6252c43a-d149-46f7-ac6c-263c34980fe2","Type":"ContainerDied","Data":"2af57a8d89d834477d936f1accea68294ed6669b9cde85ec3a277a286c70ebc0"} Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.055503 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-wst2s_a9ed1e0e-eff9-4690-bcf5-45f6074c200e/console/0.log" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.055568 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wst2s" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.184618 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8m6fm\" (UniqueName: \"kubernetes.io/projected/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-kube-api-access-8m6fm\") pod \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.184680 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-service-ca\") pod \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.184740 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-oauth-serving-cert\") pod \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.184771 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-console-oauth-config\") pod \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.184823 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-console-serving-cert\") pod \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.184882 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-trusted-ca-bundle\") pod \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.184905 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-console-config\") pod \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\" (UID: \"a9ed1e0e-eff9-4690-bcf5-45f6074c200e\") " Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.185815 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "a9ed1e0e-eff9-4690-bcf5-45f6074c200e" (UID: "a9ed1e0e-eff9-4690-bcf5-45f6074c200e"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.185826 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-service-ca" (OuterVolumeSpecName: "service-ca") pod "a9ed1e0e-eff9-4690-bcf5-45f6074c200e" (UID: "a9ed1e0e-eff9-4690-bcf5-45f6074c200e"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.185841 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-console-config" (OuterVolumeSpecName: "console-config") pod "a9ed1e0e-eff9-4690-bcf5-45f6074c200e" (UID: "a9ed1e0e-eff9-4690-bcf5-45f6074c200e"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.186206 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "a9ed1e0e-eff9-4690-bcf5-45f6074c200e" (UID: "a9ed1e0e-eff9-4690-bcf5-45f6074c200e"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.191356 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-kube-api-access-8m6fm" (OuterVolumeSpecName: "kube-api-access-8m6fm") pod "a9ed1e0e-eff9-4690-bcf5-45f6074c200e" (UID: "a9ed1e0e-eff9-4690-bcf5-45f6074c200e"). InnerVolumeSpecName "kube-api-access-8m6fm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.191524 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "a9ed1e0e-eff9-4690-bcf5-45f6074c200e" (UID: "a9ed1e0e-eff9-4690-bcf5-45f6074c200e"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.196515 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "a9ed1e0e-eff9-4690-bcf5-45f6074c200e" (UID: "a9ed1e0e-eff9-4690-bcf5-45f6074c200e"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.287471 4775 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.287565 4775 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.287587 4775 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.287604 4775 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-console-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.287622 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.287638 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8m6fm\" (UniqueName: \"kubernetes.io/projected/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-kube-api-access-8m6fm\") on node \"crc\" DevicePath \"\"" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.287658 4775 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a9ed1e0e-eff9-4690-bcf5-45f6074c200e-service-ca\") on node \"crc\" DevicePath \"\"" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.618772 4775 generic.go:334] "Generic (PLEG): container finished" podID="e3a4716c-b33b-4030-a4fa-0feda18a8657" containerID="bfefe0cac35a2f3c7e6293ed113af6198ac9e4db485c73a73295e5e65ba7c34b" exitCode=0 Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.618847 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5l6s" event={"ID":"e3a4716c-b33b-4030-a4fa-0feda18a8657","Type":"ContainerDied","Data":"bfefe0cac35a2f3c7e6293ed113af6198ac9e4db485c73a73295e5e65ba7c34b"} Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.621709 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-wst2s_a9ed1e0e-eff9-4690-bcf5-45f6074c200e/console/0.log" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.621740 4775 generic.go:334] "Generic (PLEG): container finished" podID="a9ed1e0e-eff9-4690-bcf5-45f6074c200e" containerID="4843fbb3995c43ae7854ec367e570829c36ae201e54901d87c5c625dcdce3764" exitCode=2 Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.621777 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wst2s" event={"ID":"a9ed1e0e-eff9-4690-bcf5-45f6074c200e","Type":"ContainerDied","Data":"4843fbb3995c43ae7854ec367e570829c36ae201e54901d87c5c625dcdce3764"} Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.621793 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wst2s" event={"ID":"a9ed1e0e-eff9-4690-bcf5-45f6074c200e","Type":"ContainerDied","Data":"c582e2fab200954a9843b0514b416a8988bf80f007355b323c81a87c8930c6fa"} Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.621809 4775 scope.go:117] "RemoveContainer" containerID="4843fbb3995c43ae7854ec367e570829c36ae201e54901d87c5c625dcdce3764" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.621879 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wst2s" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.627075 4775 generic.go:334] "Generic (PLEG): container finished" podID="6252c43a-d149-46f7-ac6c-263c34980fe2" containerID="a8c0ef7d7f4885716379df92240c87e2eb767b281e561cce6a2be3b2c742f219" exitCode=0 Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.627247 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s" event={"ID":"6252c43a-d149-46f7-ac6c-263c34980fe2","Type":"ContainerDied","Data":"a8c0ef7d7f4885716379df92240c87e2eb767b281e561cce6a2be3b2c742f219"} Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.641435 4775 scope.go:117] "RemoveContainer" containerID="4843fbb3995c43ae7854ec367e570829c36ae201e54901d87c5c625dcdce3764" Mar 21 05:01:30 crc kubenswrapper[4775]: E0321 05:01:30.641949 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4843fbb3995c43ae7854ec367e570829c36ae201e54901d87c5c625dcdce3764\": container with ID starting with 4843fbb3995c43ae7854ec367e570829c36ae201e54901d87c5c625dcdce3764 not found: ID does not exist" containerID="4843fbb3995c43ae7854ec367e570829c36ae201e54901d87c5c625dcdce3764" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.642026 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4843fbb3995c43ae7854ec367e570829c36ae201e54901d87c5c625dcdce3764"} err="failed to get container status \"4843fbb3995c43ae7854ec367e570829c36ae201e54901d87c5c625dcdce3764\": rpc error: code = NotFound desc = could not find container \"4843fbb3995c43ae7854ec367e570829c36ae201e54901d87c5c625dcdce3764\": container with ID starting with 4843fbb3995c43ae7854ec367e570829c36ae201e54901d87c5c625dcdce3764 not found: ID does not exist" Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.687604 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-wst2s"] Mar 21 05:01:30 crc kubenswrapper[4775]: I0321 05:01:30.693231 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-wst2s"] Mar 21 05:01:31 crc kubenswrapper[4775]: I0321 05:01:31.635605 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5l6s" event={"ID":"e3a4716c-b33b-4030-a4fa-0feda18a8657","Type":"ContainerStarted","Data":"66bada73977196eaa9ea4d56550148a1585d2a3a1a5efa78de5403ca207a8ebd"} Mar 21 05:01:31 crc kubenswrapper[4775]: I0321 05:01:31.663870 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w5l6s" podStartSLOduration=1.943428644 podStartE2EDuration="4.663843129s" podCreationTimestamp="2026-03-21 05:01:27 +0000 UTC" firstStartedPulling="2026-03-21 05:01:28.595073263 +0000 UTC m=+841.571536887" lastFinishedPulling="2026-03-21 05:01:31.315487738 +0000 UTC m=+844.291951372" observedRunningTime="2026-03-21 05:01:31.657472319 +0000 UTC m=+844.633935943" watchObservedRunningTime="2026-03-21 05:01:31.663843129 +0000 UTC m=+844.640306773" Mar 21 05:01:31 crc kubenswrapper[4775]: I0321 05:01:31.676412 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9ed1e0e-eff9-4690-bcf5-45f6074c200e" path="/var/lib/kubelet/pods/a9ed1e0e-eff9-4690-bcf5-45f6074c200e/volumes" Mar 21 05:01:31 crc kubenswrapper[4775]: I0321 05:01:31.850647 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s" Mar 21 05:01:32 crc kubenswrapper[4775]: I0321 05:01:32.006762 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6252c43a-d149-46f7-ac6c-263c34980fe2-util\") pod \"6252c43a-d149-46f7-ac6c-263c34980fe2\" (UID: \"6252c43a-d149-46f7-ac6c-263c34980fe2\") " Mar 21 05:01:32 crc kubenswrapper[4775]: I0321 05:01:32.006833 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqtp7\" (UniqueName: \"kubernetes.io/projected/6252c43a-d149-46f7-ac6c-263c34980fe2-kube-api-access-cqtp7\") pod \"6252c43a-d149-46f7-ac6c-263c34980fe2\" (UID: \"6252c43a-d149-46f7-ac6c-263c34980fe2\") " Mar 21 05:01:32 crc kubenswrapper[4775]: I0321 05:01:32.006876 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6252c43a-d149-46f7-ac6c-263c34980fe2-bundle\") pod \"6252c43a-d149-46f7-ac6c-263c34980fe2\" (UID: \"6252c43a-d149-46f7-ac6c-263c34980fe2\") " Mar 21 05:01:32 crc kubenswrapper[4775]: I0321 05:01:32.009405 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6252c43a-d149-46f7-ac6c-263c34980fe2-bundle" (OuterVolumeSpecName: "bundle") pod "6252c43a-d149-46f7-ac6c-263c34980fe2" (UID: "6252c43a-d149-46f7-ac6c-263c34980fe2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:01:32 crc kubenswrapper[4775]: I0321 05:01:32.012666 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6252c43a-d149-46f7-ac6c-263c34980fe2-kube-api-access-cqtp7" (OuterVolumeSpecName: "kube-api-access-cqtp7") pod "6252c43a-d149-46f7-ac6c-263c34980fe2" (UID: "6252c43a-d149-46f7-ac6c-263c34980fe2"). InnerVolumeSpecName "kube-api-access-cqtp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:01:32 crc kubenswrapper[4775]: I0321 05:01:32.027092 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6252c43a-d149-46f7-ac6c-263c34980fe2-util" (OuterVolumeSpecName: "util") pod "6252c43a-d149-46f7-ac6c-263c34980fe2" (UID: "6252c43a-d149-46f7-ac6c-263c34980fe2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:01:32 crc kubenswrapper[4775]: I0321 05:01:32.108301 4775 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6252c43a-d149-46f7-ac6c-263c34980fe2-util\") on node \"crc\" DevicePath \"\"" Mar 21 05:01:32 crc kubenswrapper[4775]: I0321 05:01:32.108357 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqtp7\" (UniqueName: \"kubernetes.io/projected/6252c43a-d149-46f7-ac6c-263c34980fe2-kube-api-access-cqtp7\") on node \"crc\" DevicePath \"\"" Mar 21 05:01:32 crc kubenswrapper[4775]: I0321 05:01:32.108373 4775 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6252c43a-d149-46f7-ac6c-263c34980fe2-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:01:32 crc kubenswrapper[4775]: I0321 05:01:32.482224 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:01:32 crc kubenswrapper[4775]: I0321 05:01:32.482287 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:01:32 crc kubenswrapper[4775]: I0321 05:01:32.482335 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 05:01:32 crc kubenswrapper[4775]: I0321 05:01:32.483008 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eaf646a6237d4b4ce6a8a82755505f01a336cf93fef407f02dbf82d68f5008b4"} pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 05:01:32 crc kubenswrapper[4775]: I0321 05:01:32.483091 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" containerID="cri-o://eaf646a6237d4b4ce6a8a82755505f01a336cf93fef407f02dbf82d68f5008b4" gracePeriod=600 Mar 21 05:01:32 crc kubenswrapper[4775]: I0321 05:01:32.658630 4775 generic.go:334] "Generic (PLEG): container finished" podID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerID="eaf646a6237d4b4ce6a8a82755505f01a336cf93fef407f02dbf82d68f5008b4" exitCode=0 Mar 21 05:01:32 crc kubenswrapper[4775]: I0321 05:01:32.658741 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerDied","Data":"eaf646a6237d4b4ce6a8a82755505f01a336cf93fef407f02dbf82d68f5008b4"} Mar 21 05:01:32 crc kubenswrapper[4775]: I0321 05:01:32.658800 4775 scope.go:117] "RemoveContainer" containerID="948421a752c90ac0f2fbc508e5358894c1d0ccf82922efe510f0505bdc2c2715" Mar 21 05:01:32 crc kubenswrapper[4775]: I0321 05:01:32.662900 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s" Mar 21 05:01:32 crc kubenswrapper[4775]: I0321 05:01:32.663278 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s" event={"ID":"6252c43a-d149-46f7-ac6c-263c34980fe2","Type":"ContainerDied","Data":"9848d056d20a28e5d15c2a52eb1b9b4a66212a967ad7a7690bd7227caa27e415"} Mar 21 05:01:32 crc kubenswrapper[4775]: I0321 05:01:32.663325 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9848d056d20a28e5d15c2a52eb1b9b4a66212a967ad7a7690bd7227caa27e415" Mar 21 05:01:33 crc kubenswrapper[4775]: I0321 05:01:33.694024 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerStarted","Data":"8d303e1558b2adccb16582693a30248fb7f96ca561d7dcb3104197e825dd15a7"} Mar 21 05:01:37 crc kubenswrapper[4775]: I0321 05:01:37.827841 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w5l6s" Mar 21 05:01:37 crc kubenswrapper[4775]: I0321 05:01:37.828519 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w5l6s" Mar 21 05:01:39 crc kubenswrapper[4775]: I0321 05:01:39.065857 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-w5l6s" podUID="e3a4716c-b33b-4030-a4fa-0feda18a8657" containerName="registry-server" probeResult="failure" output=< Mar 21 05:01:39 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Mar 21 05:01:39 crc kubenswrapper[4775]: > Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.570500 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6695f56dbb-f6tqn"] Mar 21 05:01:43 crc kubenswrapper[4775]: E0321 05:01:43.571305 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6252c43a-d149-46f7-ac6c-263c34980fe2" containerName="util" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.571323 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6252c43a-d149-46f7-ac6c-263c34980fe2" containerName="util" Mar 21 05:01:43 crc kubenswrapper[4775]: E0321 05:01:43.571347 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ed1e0e-eff9-4690-bcf5-45f6074c200e" containerName="console" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.571354 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ed1e0e-eff9-4690-bcf5-45f6074c200e" containerName="console" Mar 21 05:01:43 crc kubenswrapper[4775]: E0321 05:01:43.571368 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6252c43a-d149-46f7-ac6c-263c34980fe2" containerName="pull" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.571376 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6252c43a-d149-46f7-ac6c-263c34980fe2" containerName="pull" Mar 21 05:01:43 crc kubenswrapper[4775]: E0321 05:01:43.571414 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6252c43a-d149-46f7-ac6c-263c34980fe2" containerName="extract" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.571421 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6252c43a-d149-46f7-ac6c-263c34980fe2" containerName="extract" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.571543 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9ed1e0e-eff9-4690-bcf5-45f6074c200e" containerName="console" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.571567 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6252c43a-d149-46f7-ac6c-263c34980fe2" containerName="extract" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.572021 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6695f56dbb-f6tqn" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.576529 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.576841 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.577260 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.579322 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.581162 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-msjcl" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.612309 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6695f56dbb-f6tqn"] Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.619425 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/38d80f78-fa33-49b7-99c2-62d50d1c011b-apiservice-cert\") pod \"metallb-operator-controller-manager-6695f56dbb-f6tqn\" (UID: \"38d80f78-fa33-49b7-99c2-62d50d1c011b\") " pod="metallb-system/metallb-operator-controller-manager-6695f56dbb-f6tqn" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.619500 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/38d80f78-fa33-49b7-99c2-62d50d1c011b-webhook-cert\") pod \"metallb-operator-controller-manager-6695f56dbb-f6tqn\" (UID: \"38d80f78-fa33-49b7-99c2-62d50d1c011b\") " pod="metallb-system/metallb-operator-controller-manager-6695f56dbb-f6tqn" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.619540 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb947\" (UniqueName: \"kubernetes.io/projected/38d80f78-fa33-49b7-99c2-62d50d1c011b-kube-api-access-hb947\") pod \"metallb-operator-controller-manager-6695f56dbb-f6tqn\" (UID: \"38d80f78-fa33-49b7-99c2-62d50d1c011b\") " pod="metallb-system/metallb-operator-controller-manager-6695f56dbb-f6tqn" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.726665 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/38d80f78-fa33-49b7-99c2-62d50d1c011b-webhook-cert\") pod \"metallb-operator-controller-manager-6695f56dbb-f6tqn\" (UID: \"38d80f78-fa33-49b7-99c2-62d50d1c011b\") " pod="metallb-system/metallb-operator-controller-manager-6695f56dbb-f6tqn" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.726723 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb947\" (UniqueName: \"kubernetes.io/projected/38d80f78-fa33-49b7-99c2-62d50d1c011b-kube-api-access-hb947\") pod \"metallb-operator-controller-manager-6695f56dbb-f6tqn\" (UID: \"38d80f78-fa33-49b7-99c2-62d50d1c011b\") " pod="metallb-system/metallb-operator-controller-manager-6695f56dbb-f6tqn" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.726862 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/38d80f78-fa33-49b7-99c2-62d50d1c011b-apiservice-cert\") pod \"metallb-operator-controller-manager-6695f56dbb-f6tqn\" (UID: \"38d80f78-fa33-49b7-99c2-62d50d1c011b\") " pod="metallb-system/metallb-operator-controller-manager-6695f56dbb-f6tqn" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.736821 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/38d80f78-fa33-49b7-99c2-62d50d1c011b-apiservice-cert\") pod \"metallb-operator-controller-manager-6695f56dbb-f6tqn\" (UID: \"38d80f78-fa33-49b7-99c2-62d50d1c011b\") " pod="metallb-system/metallb-operator-controller-manager-6695f56dbb-f6tqn" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.738052 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/38d80f78-fa33-49b7-99c2-62d50d1c011b-webhook-cert\") pod \"metallb-operator-controller-manager-6695f56dbb-f6tqn\" (UID: \"38d80f78-fa33-49b7-99c2-62d50d1c011b\") " pod="metallb-system/metallb-operator-controller-manager-6695f56dbb-f6tqn" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.795805 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb947\" (UniqueName: \"kubernetes.io/projected/38d80f78-fa33-49b7-99c2-62d50d1c011b-kube-api-access-hb947\") pod \"metallb-operator-controller-manager-6695f56dbb-f6tqn\" (UID: \"38d80f78-fa33-49b7-99c2-62d50d1c011b\") " pod="metallb-system/metallb-operator-controller-manager-6695f56dbb-f6tqn" Mar 21 05:01:43 crc kubenswrapper[4775]: I0321 05:01:43.893590 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6695f56dbb-f6tqn" Mar 21 05:01:44 crc kubenswrapper[4775]: I0321 05:01:44.026564 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-559bfcf5c-qqsvn"] Mar 21 05:01:44 crc kubenswrapper[4775]: I0321 05:01:44.027273 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-559bfcf5c-qqsvn" Mar 21 05:01:44 crc kubenswrapper[4775]: I0321 05:01:44.031159 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 21 05:01:44 crc kubenswrapper[4775]: I0321 05:01:44.031480 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-84rwk" Mar 21 05:01:44 crc kubenswrapper[4775]: I0321 05:01:44.034980 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 21 05:01:44 crc kubenswrapper[4775]: I0321 05:01:44.052883 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-559bfcf5c-qqsvn"] Mar 21 05:01:44 crc kubenswrapper[4775]: I0321 05:01:44.133018 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jrhz\" (UniqueName: \"kubernetes.io/projected/8c52832d-1aee-4eac-b625-24110b985402-kube-api-access-4jrhz\") pod \"metallb-operator-webhook-server-559bfcf5c-qqsvn\" (UID: \"8c52832d-1aee-4eac-b625-24110b985402\") " pod="metallb-system/metallb-operator-webhook-server-559bfcf5c-qqsvn" Mar 21 05:01:44 crc kubenswrapper[4775]: I0321 05:01:44.133439 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8c52832d-1aee-4eac-b625-24110b985402-apiservice-cert\") pod \"metallb-operator-webhook-server-559bfcf5c-qqsvn\" (UID: \"8c52832d-1aee-4eac-b625-24110b985402\") " pod="metallb-system/metallb-operator-webhook-server-559bfcf5c-qqsvn" Mar 21 05:01:44 crc kubenswrapper[4775]: I0321 05:01:44.133481 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8c52832d-1aee-4eac-b625-24110b985402-webhook-cert\") pod \"metallb-operator-webhook-server-559bfcf5c-qqsvn\" (UID: \"8c52832d-1aee-4eac-b625-24110b985402\") " pod="metallb-system/metallb-operator-webhook-server-559bfcf5c-qqsvn" Mar 21 05:01:44 crc kubenswrapper[4775]: I0321 05:01:44.234232 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jrhz\" (UniqueName: \"kubernetes.io/projected/8c52832d-1aee-4eac-b625-24110b985402-kube-api-access-4jrhz\") pod \"metallb-operator-webhook-server-559bfcf5c-qqsvn\" (UID: \"8c52832d-1aee-4eac-b625-24110b985402\") " pod="metallb-system/metallb-operator-webhook-server-559bfcf5c-qqsvn" Mar 21 05:01:44 crc kubenswrapper[4775]: I0321 05:01:44.234275 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8c52832d-1aee-4eac-b625-24110b985402-apiservice-cert\") pod \"metallb-operator-webhook-server-559bfcf5c-qqsvn\" (UID: \"8c52832d-1aee-4eac-b625-24110b985402\") " pod="metallb-system/metallb-operator-webhook-server-559bfcf5c-qqsvn" Mar 21 05:01:44 crc kubenswrapper[4775]: I0321 05:01:44.234312 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8c52832d-1aee-4eac-b625-24110b985402-webhook-cert\") pod \"metallb-operator-webhook-server-559bfcf5c-qqsvn\" (UID: \"8c52832d-1aee-4eac-b625-24110b985402\") " pod="metallb-system/metallb-operator-webhook-server-559bfcf5c-qqsvn" Mar 21 05:01:44 crc kubenswrapper[4775]: I0321 05:01:44.240826 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8c52832d-1aee-4eac-b625-24110b985402-webhook-cert\") pod \"metallb-operator-webhook-server-559bfcf5c-qqsvn\" (UID: \"8c52832d-1aee-4eac-b625-24110b985402\") " pod="metallb-system/metallb-operator-webhook-server-559bfcf5c-qqsvn" Mar 21 05:01:44 crc kubenswrapper[4775]: I0321 05:01:44.241869 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8c52832d-1aee-4eac-b625-24110b985402-apiservice-cert\") pod \"metallb-operator-webhook-server-559bfcf5c-qqsvn\" (UID: \"8c52832d-1aee-4eac-b625-24110b985402\") " pod="metallb-system/metallb-operator-webhook-server-559bfcf5c-qqsvn" Mar 21 05:01:44 crc kubenswrapper[4775]: I0321 05:01:44.255946 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jrhz\" (UniqueName: \"kubernetes.io/projected/8c52832d-1aee-4eac-b625-24110b985402-kube-api-access-4jrhz\") pod \"metallb-operator-webhook-server-559bfcf5c-qqsvn\" (UID: \"8c52832d-1aee-4eac-b625-24110b985402\") " pod="metallb-system/metallb-operator-webhook-server-559bfcf5c-qqsvn" Mar 21 05:01:44 crc kubenswrapper[4775]: I0321 05:01:44.351885 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-559bfcf5c-qqsvn" Mar 21 05:01:44 crc kubenswrapper[4775]: I0321 05:01:44.356639 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6695f56dbb-f6tqn"] Mar 21 05:01:44 crc kubenswrapper[4775]: W0321 05:01:44.387186 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38d80f78_fa33_49b7_99c2_62d50d1c011b.slice/crio-fdd55ebd093a8f7c065a1d3ab318dd78e829212ce1f2638f9ee9aabd5cffde3f WatchSource:0}: Error finding container fdd55ebd093a8f7c065a1d3ab318dd78e829212ce1f2638f9ee9aabd5cffde3f: Status 404 returned error can't find the container with id fdd55ebd093a8f7c065a1d3ab318dd78e829212ce1f2638f9ee9aabd5cffde3f Mar 21 05:01:44 crc kubenswrapper[4775]: I0321 05:01:44.565950 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-559bfcf5c-qqsvn"] Mar 21 05:01:44 crc kubenswrapper[4775]: W0321 05:01:44.577328 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c52832d_1aee_4eac_b625_24110b985402.slice/crio-c9a74335940a7c570ab8a8fdf97fcea4e88cf8d6d72777f0d31424d8d6bebd67 WatchSource:0}: Error finding container c9a74335940a7c570ab8a8fdf97fcea4e88cf8d6d72777f0d31424d8d6bebd67: Status 404 returned error can't find the container with id c9a74335940a7c570ab8a8fdf97fcea4e88cf8d6d72777f0d31424d8d6bebd67 Mar 21 05:01:44 crc kubenswrapper[4775]: I0321 05:01:44.766083 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6695f56dbb-f6tqn" event={"ID":"38d80f78-fa33-49b7-99c2-62d50d1c011b","Type":"ContainerStarted","Data":"fdd55ebd093a8f7c065a1d3ab318dd78e829212ce1f2638f9ee9aabd5cffde3f"} Mar 21 05:01:44 crc kubenswrapper[4775]: I0321 05:01:44.768622 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-559bfcf5c-qqsvn" event={"ID":"8c52832d-1aee-4eac-b625-24110b985402","Type":"ContainerStarted","Data":"c9a74335940a7c570ab8a8fdf97fcea4e88cf8d6d72777f0d31424d8d6bebd67"} Mar 21 05:01:47 crc kubenswrapper[4775]: I0321 05:01:47.879796 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w5l6s" Mar 21 05:01:47 crc kubenswrapper[4775]: I0321 05:01:47.928983 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w5l6s" Mar 21 05:01:48 crc kubenswrapper[4775]: I0321 05:01:48.690856 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w5l6s"] Mar 21 05:01:49 crc kubenswrapper[4775]: I0321 05:01:49.820582 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w5l6s" podUID="e3a4716c-b33b-4030-a4fa-0feda18a8657" containerName="registry-server" containerID="cri-o://66bada73977196eaa9ea4d56550148a1585d2a3a1a5efa78de5403ca207a8ebd" gracePeriod=2 Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.475415 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w5l6s" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.562702 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ww7q\" (UniqueName: \"kubernetes.io/projected/e3a4716c-b33b-4030-a4fa-0feda18a8657-kube-api-access-5ww7q\") pod \"e3a4716c-b33b-4030-a4fa-0feda18a8657\" (UID: \"e3a4716c-b33b-4030-a4fa-0feda18a8657\") " Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.562789 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3a4716c-b33b-4030-a4fa-0feda18a8657-catalog-content\") pod \"e3a4716c-b33b-4030-a4fa-0feda18a8657\" (UID: \"e3a4716c-b33b-4030-a4fa-0feda18a8657\") " Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.562820 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3a4716c-b33b-4030-a4fa-0feda18a8657-utilities\") pod \"e3a4716c-b33b-4030-a4fa-0feda18a8657\" (UID: \"e3a4716c-b33b-4030-a4fa-0feda18a8657\") " Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.563945 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3a4716c-b33b-4030-a4fa-0feda18a8657-utilities" (OuterVolumeSpecName: "utilities") pod "e3a4716c-b33b-4030-a4fa-0feda18a8657" (UID: "e3a4716c-b33b-4030-a4fa-0feda18a8657"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.573392 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3a4716c-b33b-4030-a4fa-0feda18a8657-kube-api-access-5ww7q" (OuterVolumeSpecName: "kube-api-access-5ww7q") pod "e3a4716c-b33b-4030-a4fa-0feda18a8657" (UID: "e3a4716c-b33b-4030-a4fa-0feda18a8657"). InnerVolumeSpecName "kube-api-access-5ww7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.664243 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ww7q\" (UniqueName: \"kubernetes.io/projected/e3a4716c-b33b-4030-a4fa-0feda18a8657-kube-api-access-5ww7q\") on node \"crc\" DevicePath \"\"" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.664576 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3a4716c-b33b-4030-a4fa-0feda18a8657-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.709863 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3a4716c-b33b-4030-a4fa-0feda18a8657-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e3a4716c-b33b-4030-a4fa-0feda18a8657" (UID: "e3a4716c-b33b-4030-a4fa-0feda18a8657"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.766183 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3a4716c-b33b-4030-a4fa-0feda18a8657-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.827025 4775 generic.go:334] "Generic (PLEG): container finished" podID="e3a4716c-b33b-4030-a4fa-0feda18a8657" containerID="66bada73977196eaa9ea4d56550148a1585d2a3a1a5efa78de5403ca207a8ebd" exitCode=0 Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.827081 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5l6s" event={"ID":"e3a4716c-b33b-4030-a4fa-0feda18a8657","Type":"ContainerDied","Data":"66bada73977196eaa9ea4d56550148a1585d2a3a1a5efa78de5403ca207a8ebd"} Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.827110 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5l6s" event={"ID":"e3a4716c-b33b-4030-a4fa-0feda18a8657","Type":"ContainerDied","Data":"e0c863098fc3ee0671df4688c9648d37c09a40f58d77f06335e48b368ac9c588"} Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.827160 4775 scope.go:117] "RemoveContainer" containerID="66bada73977196eaa9ea4d56550148a1585d2a3a1a5efa78de5403ca207a8ebd" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.827301 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w5l6s" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.834346 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6695f56dbb-f6tqn" event={"ID":"38d80f78-fa33-49b7-99c2-62d50d1c011b","Type":"ContainerStarted","Data":"0828245983a7e49a9e117cc12dfaf0365eeb35c9e1e14cb85941fcf0da8e81b5"} Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.834962 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6695f56dbb-f6tqn" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.836750 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-559bfcf5c-qqsvn" event={"ID":"8c52832d-1aee-4eac-b625-24110b985402","Type":"ContainerStarted","Data":"22e178e2bce37747bac8ef46e8573736db1a74306e9f3e342f7fee33b4fe30f2"} Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.837217 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-559bfcf5c-qqsvn" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.855869 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6695f56dbb-f6tqn" podStartSLOduration=2.207598325 podStartE2EDuration="7.855847893s" podCreationTimestamp="2026-03-21 05:01:43 +0000 UTC" firstStartedPulling="2026-03-21 05:01:44.391714415 +0000 UTC m=+857.368178039" lastFinishedPulling="2026-03-21 05:01:50.039963983 +0000 UTC m=+863.016427607" observedRunningTime="2026-03-21 05:01:50.853221689 +0000 UTC m=+863.829685313" watchObservedRunningTime="2026-03-21 05:01:50.855847893 +0000 UTC m=+863.832311517" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.858697 4775 scope.go:117] "RemoveContainer" containerID="bfefe0cac35a2f3c7e6293ed113af6198ac9e4db485c73a73295e5e65ba7c34b" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.876854 4775 scope.go:117] "RemoveContainer" containerID="2adc93f631ab51a3bffbfd6d7c28133602106d191f43a08a57386749aee6506a" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.877968 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-559bfcf5c-qqsvn" podStartSLOduration=1.985400447 podStartE2EDuration="7.877946847s" podCreationTimestamp="2026-03-21 05:01:43 +0000 UTC" firstStartedPulling="2026-03-21 05:01:44.57877949 +0000 UTC m=+857.555243114" lastFinishedPulling="2026-03-21 05:01:50.47132589 +0000 UTC m=+863.447789514" observedRunningTime="2026-03-21 05:01:50.874943363 +0000 UTC m=+863.851406987" watchObservedRunningTime="2026-03-21 05:01:50.877946847 +0000 UTC m=+863.854410471" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.890697 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w5l6s"] Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.894657 4775 scope.go:117] "RemoveContainer" containerID="66bada73977196eaa9ea4d56550148a1585d2a3a1a5efa78de5403ca207a8ebd" Mar 21 05:01:50 crc kubenswrapper[4775]: E0321 05:01:50.895473 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66bada73977196eaa9ea4d56550148a1585d2a3a1a5efa78de5403ca207a8ebd\": container with ID starting with 66bada73977196eaa9ea4d56550148a1585d2a3a1a5efa78de5403ca207a8ebd not found: ID does not exist" containerID="66bada73977196eaa9ea4d56550148a1585d2a3a1a5efa78de5403ca207a8ebd" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.895524 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66bada73977196eaa9ea4d56550148a1585d2a3a1a5efa78de5403ca207a8ebd"} err="failed to get container status \"66bada73977196eaa9ea4d56550148a1585d2a3a1a5efa78de5403ca207a8ebd\": rpc error: code = NotFound desc = could not find container \"66bada73977196eaa9ea4d56550148a1585d2a3a1a5efa78de5403ca207a8ebd\": container with ID starting with 66bada73977196eaa9ea4d56550148a1585d2a3a1a5efa78de5403ca207a8ebd not found: ID does not exist" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.895611 4775 scope.go:117] "RemoveContainer" containerID="bfefe0cac35a2f3c7e6293ed113af6198ac9e4db485c73a73295e5e65ba7c34b" Mar 21 05:01:50 crc kubenswrapper[4775]: E0321 05:01:50.896771 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfefe0cac35a2f3c7e6293ed113af6198ac9e4db485c73a73295e5e65ba7c34b\": container with ID starting with bfefe0cac35a2f3c7e6293ed113af6198ac9e4db485c73a73295e5e65ba7c34b not found: ID does not exist" containerID="bfefe0cac35a2f3c7e6293ed113af6198ac9e4db485c73a73295e5e65ba7c34b" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.896844 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfefe0cac35a2f3c7e6293ed113af6198ac9e4db485c73a73295e5e65ba7c34b"} err="failed to get container status \"bfefe0cac35a2f3c7e6293ed113af6198ac9e4db485c73a73295e5e65ba7c34b\": rpc error: code = NotFound desc = could not find container \"bfefe0cac35a2f3c7e6293ed113af6198ac9e4db485c73a73295e5e65ba7c34b\": container with ID starting with bfefe0cac35a2f3c7e6293ed113af6198ac9e4db485c73a73295e5e65ba7c34b not found: ID does not exist" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.896881 4775 scope.go:117] "RemoveContainer" containerID="2adc93f631ab51a3bffbfd6d7c28133602106d191f43a08a57386749aee6506a" Mar 21 05:01:50 crc kubenswrapper[4775]: E0321 05:01:50.901339 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2adc93f631ab51a3bffbfd6d7c28133602106d191f43a08a57386749aee6506a\": container with ID starting with 2adc93f631ab51a3bffbfd6d7c28133602106d191f43a08a57386749aee6506a not found: ID does not exist" containerID="2adc93f631ab51a3bffbfd6d7c28133602106d191f43a08a57386749aee6506a" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.901490 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2adc93f631ab51a3bffbfd6d7c28133602106d191f43a08a57386749aee6506a"} err="failed to get container status \"2adc93f631ab51a3bffbfd6d7c28133602106d191f43a08a57386749aee6506a\": rpc error: code = NotFound desc = could not find container \"2adc93f631ab51a3bffbfd6d7c28133602106d191f43a08a57386749aee6506a\": container with ID starting with 2adc93f631ab51a3bffbfd6d7c28133602106d191f43a08a57386749aee6506a not found: ID does not exist" Mar 21 05:01:50 crc kubenswrapper[4775]: I0321 05:01:50.904559 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w5l6s"] Mar 21 05:01:51 crc kubenswrapper[4775]: I0321 05:01:51.668710 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3a4716c-b33b-4030-a4fa-0feda18a8657" path="/var/lib/kubelet/pods/e3a4716c-b33b-4030-a4fa-0feda18a8657/volumes" Mar 21 05:02:00 crc kubenswrapper[4775]: I0321 05:02:00.153847 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567822-pqlrw"] Mar 21 05:02:00 crc kubenswrapper[4775]: E0321 05:02:00.154730 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3a4716c-b33b-4030-a4fa-0feda18a8657" containerName="registry-server" Mar 21 05:02:00 crc kubenswrapper[4775]: I0321 05:02:00.154748 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3a4716c-b33b-4030-a4fa-0feda18a8657" containerName="registry-server" Mar 21 05:02:00 crc kubenswrapper[4775]: E0321 05:02:00.154759 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3a4716c-b33b-4030-a4fa-0feda18a8657" containerName="extract-utilities" Mar 21 05:02:00 crc kubenswrapper[4775]: I0321 05:02:00.154766 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3a4716c-b33b-4030-a4fa-0feda18a8657" containerName="extract-utilities" Mar 21 05:02:00 crc kubenswrapper[4775]: E0321 05:02:00.154775 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3a4716c-b33b-4030-a4fa-0feda18a8657" containerName="extract-content" Mar 21 05:02:00 crc kubenswrapper[4775]: I0321 05:02:00.154781 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3a4716c-b33b-4030-a4fa-0feda18a8657" containerName="extract-content" Mar 21 05:02:00 crc kubenswrapper[4775]: I0321 05:02:00.154906 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3a4716c-b33b-4030-a4fa-0feda18a8657" containerName="registry-server" Mar 21 05:02:00 crc kubenswrapper[4775]: I0321 05:02:00.155353 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567822-pqlrw" Mar 21 05:02:00 crc kubenswrapper[4775]: I0321 05:02:00.157790 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:02:00 crc kubenswrapper[4775]: I0321 05:02:00.158384 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:02:00 crc kubenswrapper[4775]: I0321 05:02:00.158661 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:02:00 crc kubenswrapper[4775]: I0321 05:02:00.175221 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567822-pqlrw"] Mar 21 05:02:00 crc kubenswrapper[4775]: I0321 05:02:00.297686 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcgjc\" (UniqueName: \"kubernetes.io/projected/f791f730-ffa2-44d3-b151-dc9d6e6e1743-kube-api-access-dcgjc\") pod \"auto-csr-approver-29567822-pqlrw\" (UID: \"f791f730-ffa2-44d3-b151-dc9d6e6e1743\") " pod="openshift-infra/auto-csr-approver-29567822-pqlrw" Mar 21 05:02:00 crc kubenswrapper[4775]: I0321 05:02:00.398985 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcgjc\" (UniqueName: \"kubernetes.io/projected/f791f730-ffa2-44d3-b151-dc9d6e6e1743-kube-api-access-dcgjc\") pod \"auto-csr-approver-29567822-pqlrw\" (UID: \"f791f730-ffa2-44d3-b151-dc9d6e6e1743\") " pod="openshift-infra/auto-csr-approver-29567822-pqlrw" Mar 21 05:02:00 crc kubenswrapper[4775]: I0321 05:02:00.423795 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcgjc\" (UniqueName: \"kubernetes.io/projected/f791f730-ffa2-44d3-b151-dc9d6e6e1743-kube-api-access-dcgjc\") pod \"auto-csr-approver-29567822-pqlrw\" (UID: \"f791f730-ffa2-44d3-b151-dc9d6e6e1743\") " pod="openshift-infra/auto-csr-approver-29567822-pqlrw" Mar 21 05:02:00 crc kubenswrapper[4775]: I0321 05:02:00.472262 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567822-pqlrw" Mar 21 05:02:00 crc kubenswrapper[4775]: I0321 05:02:00.900555 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567822-pqlrw"] Mar 21 05:02:01 crc kubenswrapper[4775]: I0321 05:02:01.898770 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567822-pqlrw" event={"ID":"f791f730-ffa2-44d3-b151-dc9d6e6e1743","Type":"ContainerStarted","Data":"439adf561bf5728b29d86d131eb9f13bcfb89b7a3f50e301bdeb183340af5ce2"} Mar 21 05:02:02 crc kubenswrapper[4775]: I0321 05:02:02.907189 4775 generic.go:334] "Generic (PLEG): container finished" podID="f791f730-ffa2-44d3-b151-dc9d6e6e1743" containerID="a9b0f66cd072ab055301f7bd9fc0725dbbd203bf7fa9a48d07d610fb2904d30b" exitCode=0 Mar 21 05:02:02 crc kubenswrapper[4775]: I0321 05:02:02.907232 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567822-pqlrw" event={"ID":"f791f730-ffa2-44d3-b151-dc9d6e6e1743","Type":"ContainerDied","Data":"a9b0f66cd072ab055301f7bd9fc0725dbbd203bf7fa9a48d07d610fb2904d30b"} Mar 21 05:02:04 crc kubenswrapper[4775]: I0321 05:02:04.214516 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567822-pqlrw" Mar 21 05:02:04 crc kubenswrapper[4775]: I0321 05:02:04.265598 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcgjc\" (UniqueName: \"kubernetes.io/projected/f791f730-ffa2-44d3-b151-dc9d6e6e1743-kube-api-access-dcgjc\") pod \"f791f730-ffa2-44d3-b151-dc9d6e6e1743\" (UID: \"f791f730-ffa2-44d3-b151-dc9d6e6e1743\") " Mar 21 05:02:04 crc kubenswrapper[4775]: I0321 05:02:04.273294 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f791f730-ffa2-44d3-b151-dc9d6e6e1743-kube-api-access-dcgjc" (OuterVolumeSpecName: "kube-api-access-dcgjc") pod "f791f730-ffa2-44d3-b151-dc9d6e6e1743" (UID: "f791f730-ffa2-44d3-b151-dc9d6e6e1743"). InnerVolumeSpecName "kube-api-access-dcgjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:02:04 crc kubenswrapper[4775]: I0321 05:02:04.360589 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-559bfcf5c-qqsvn" Mar 21 05:02:04 crc kubenswrapper[4775]: I0321 05:02:04.366680 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcgjc\" (UniqueName: \"kubernetes.io/projected/f791f730-ffa2-44d3-b151-dc9d6e6e1743-kube-api-access-dcgjc\") on node \"crc\" DevicePath \"\"" Mar 21 05:02:04 crc kubenswrapper[4775]: I0321 05:02:04.919622 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567822-pqlrw" event={"ID":"f791f730-ffa2-44d3-b151-dc9d6e6e1743","Type":"ContainerDied","Data":"439adf561bf5728b29d86d131eb9f13bcfb89b7a3f50e301bdeb183340af5ce2"} Mar 21 05:02:04 crc kubenswrapper[4775]: I0321 05:02:04.919659 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="439adf561bf5728b29d86d131eb9f13bcfb89b7a3f50e301bdeb183340af5ce2" Mar 21 05:02:04 crc kubenswrapper[4775]: I0321 05:02:04.919691 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567822-pqlrw" Mar 21 05:02:05 crc kubenswrapper[4775]: I0321 05:02:05.263897 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567816-7wggn"] Mar 21 05:02:05 crc kubenswrapper[4775]: I0321 05:02:05.276322 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567816-7wggn"] Mar 21 05:02:05 crc kubenswrapper[4775]: I0321 05:02:05.555964 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hxvz8"] Mar 21 05:02:05 crc kubenswrapper[4775]: E0321 05:02:05.556255 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f791f730-ffa2-44d3-b151-dc9d6e6e1743" containerName="oc" Mar 21 05:02:05 crc kubenswrapper[4775]: I0321 05:02:05.556274 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f791f730-ffa2-44d3-b151-dc9d6e6e1743" containerName="oc" Mar 21 05:02:05 crc kubenswrapper[4775]: I0321 05:02:05.556394 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f791f730-ffa2-44d3-b151-dc9d6e6e1743" containerName="oc" Mar 21 05:02:05 crc kubenswrapper[4775]: I0321 05:02:05.557353 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hxvz8" Mar 21 05:02:05 crc kubenswrapper[4775]: I0321 05:02:05.566541 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hxvz8"] Mar 21 05:02:05 crc kubenswrapper[4775]: I0321 05:02:05.667508 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1edcd264-4604-481b-b146-1ed5a34badba" path="/var/lib/kubelet/pods/1edcd264-4604-481b-b146-1ed5a34badba/volumes" Mar 21 05:02:05 crc kubenswrapper[4775]: I0321 05:02:05.690322 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65fae7ea-6284-4905-aa68-3251c5ef0602-catalog-content\") pod \"certified-operators-hxvz8\" (UID: \"65fae7ea-6284-4905-aa68-3251c5ef0602\") " pod="openshift-marketplace/certified-operators-hxvz8" Mar 21 05:02:05 crc kubenswrapper[4775]: I0321 05:02:05.690419 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dp2c\" (UniqueName: \"kubernetes.io/projected/65fae7ea-6284-4905-aa68-3251c5ef0602-kube-api-access-5dp2c\") pod \"certified-operators-hxvz8\" (UID: \"65fae7ea-6284-4905-aa68-3251c5ef0602\") " pod="openshift-marketplace/certified-operators-hxvz8" Mar 21 05:02:05 crc kubenswrapper[4775]: I0321 05:02:05.690477 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65fae7ea-6284-4905-aa68-3251c5ef0602-utilities\") pod \"certified-operators-hxvz8\" (UID: \"65fae7ea-6284-4905-aa68-3251c5ef0602\") " pod="openshift-marketplace/certified-operators-hxvz8" Mar 21 05:02:05 crc kubenswrapper[4775]: I0321 05:02:05.791468 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65fae7ea-6284-4905-aa68-3251c5ef0602-utilities\") pod \"certified-operators-hxvz8\" (UID: \"65fae7ea-6284-4905-aa68-3251c5ef0602\") " pod="openshift-marketplace/certified-operators-hxvz8" Mar 21 05:02:05 crc kubenswrapper[4775]: I0321 05:02:05.791730 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65fae7ea-6284-4905-aa68-3251c5ef0602-catalog-content\") pod \"certified-operators-hxvz8\" (UID: \"65fae7ea-6284-4905-aa68-3251c5ef0602\") " pod="openshift-marketplace/certified-operators-hxvz8" Mar 21 05:02:05 crc kubenswrapper[4775]: I0321 05:02:05.791803 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dp2c\" (UniqueName: \"kubernetes.io/projected/65fae7ea-6284-4905-aa68-3251c5ef0602-kube-api-access-5dp2c\") pod \"certified-operators-hxvz8\" (UID: \"65fae7ea-6284-4905-aa68-3251c5ef0602\") " pod="openshift-marketplace/certified-operators-hxvz8" Mar 21 05:02:05 crc kubenswrapper[4775]: I0321 05:02:05.792110 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65fae7ea-6284-4905-aa68-3251c5ef0602-utilities\") pod \"certified-operators-hxvz8\" (UID: \"65fae7ea-6284-4905-aa68-3251c5ef0602\") " pod="openshift-marketplace/certified-operators-hxvz8" Mar 21 05:02:05 crc kubenswrapper[4775]: I0321 05:02:05.792250 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65fae7ea-6284-4905-aa68-3251c5ef0602-catalog-content\") pod \"certified-operators-hxvz8\" (UID: \"65fae7ea-6284-4905-aa68-3251c5ef0602\") " pod="openshift-marketplace/certified-operators-hxvz8" Mar 21 05:02:05 crc kubenswrapper[4775]: I0321 05:02:05.811363 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dp2c\" (UniqueName: \"kubernetes.io/projected/65fae7ea-6284-4905-aa68-3251c5ef0602-kube-api-access-5dp2c\") pod \"certified-operators-hxvz8\" (UID: \"65fae7ea-6284-4905-aa68-3251c5ef0602\") " pod="openshift-marketplace/certified-operators-hxvz8" Mar 21 05:02:05 crc kubenswrapper[4775]: I0321 05:02:05.874094 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hxvz8" Mar 21 05:02:06 crc kubenswrapper[4775]: I0321 05:02:06.164633 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hxvz8"] Mar 21 05:02:06 crc kubenswrapper[4775]: I0321 05:02:06.947654 4775 generic.go:334] "Generic (PLEG): container finished" podID="65fae7ea-6284-4905-aa68-3251c5ef0602" containerID="cc6575c3206e4931e90f2d2eeedfd610fe0c60e3f48437becf4f5af661bb2558" exitCode=0 Mar 21 05:02:06 crc kubenswrapper[4775]: I0321 05:02:06.947726 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hxvz8" event={"ID":"65fae7ea-6284-4905-aa68-3251c5ef0602","Type":"ContainerDied","Data":"cc6575c3206e4931e90f2d2eeedfd610fe0c60e3f48437becf4f5af661bb2558"} Mar 21 05:02:06 crc kubenswrapper[4775]: I0321 05:02:06.947968 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hxvz8" event={"ID":"65fae7ea-6284-4905-aa68-3251c5ef0602","Type":"ContainerStarted","Data":"b496a0cb17e25c9afc1e8fbd644c8d6a7c4698da16d39e00bebd370b5d3a0cc3"} Mar 21 05:02:08 crc kubenswrapper[4775]: I0321 05:02:08.962301 4775 generic.go:334] "Generic (PLEG): container finished" podID="65fae7ea-6284-4905-aa68-3251c5ef0602" containerID="e6b7d92e6c0e684618dc387b9902e9548ae9e2cf09c7362448b678bb11ed2975" exitCode=0 Mar 21 05:02:08 crc kubenswrapper[4775]: I0321 05:02:08.962404 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hxvz8" event={"ID":"65fae7ea-6284-4905-aa68-3251c5ef0602","Type":"ContainerDied","Data":"e6b7d92e6c0e684618dc387b9902e9548ae9e2cf09c7362448b678bb11ed2975"} Mar 21 05:02:09 crc kubenswrapper[4775]: I0321 05:02:09.970871 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hxvz8" event={"ID":"65fae7ea-6284-4905-aa68-3251c5ef0602","Type":"ContainerStarted","Data":"9cec0f381408ed32105d216b30f0489513e66921ad73f2364aedcec396fd2f22"} Mar 21 05:02:09 crc kubenswrapper[4775]: I0321 05:02:09.993300 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hxvz8" podStartSLOduration=2.155120302 podStartE2EDuration="4.993282804s" podCreationTimestamp="2026-03-21 05:02:05 +0000 UTC" firstStartedPulling="2026-03-21 05:02:06.950329678 +0000 UTC m=+879.926793342" lastFinishedPulling="2026-03-21 05:02:09.78849223 +0000 UTC m=+882.764955844" observedRunningTime="2026-03-21 05:02:09.989981521 +0000 UTC m=+882.966445165" watchObservedRunningTime="2026-03-21 05:02:09.993282804 +0000 UTC m=+882.969746428" Mar 21 05:02:15 crc kubenswrapper[4775]: I0321 05:02:15.874363 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hxvz8" Mar 21 05:02:15 crc kubenswrapper[4775]: I0321 05:02:15.875096 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hxvz8" Mar 21 05:02:15 crc kubenswrapper[4775]: I0321 05:02:15.936732 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hxvz8" Mar 21 05:02:16 crc kubenswrapper[4775]: I0321 05:02:16.044647 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hxvz8" Mar 21 05:02:18 crc kubenswrapper[4775]: I0321 05:02:18.346651 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hxvz8"] Mar 21 05:02:19 crc kubenswrapper[4775]: I0321 05:02:19.023698 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hxvz8" podUID="65fae7ea-6284-4905-aa68-3251c5ef0602" containerName="registry-server" containerID="cri-o://9cec0f381408ed32105d216b30f0489513e66921ad73f2364aedcec396fd2f22" gracePeriod=2 Mar 21 05:02:21 crc kubenswrapper[4775]: I0321 05:02:21.039144 4775 generic.go:334] "Generic (PLEG): container finished" podID="65fae7ea-6284-4905-aa68-3251c5ef0602" containerID="9cec0f381408ed32105d216b30f0489513e66921ad73f2364aedcec396fd2f22" exitCode=0 Mar 21 05:02:21 crc kubenswrapper[4775]: I0321 05:02:21.039157 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hxvz8" event={"ID":"65fae7ea-6284-4905-aa68-3251c5ef0602","Type":"ContainerDied","Data":"9cec0f381408ed32105d216b30f0489513e66921ad73f2364aedcec396fd2f22"} Mar 21 05:02:21 crc kubenswrapper[4775]: I0321 05:02:21.474747 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hxvz8" Mar 21 05:02:21 crc kubenswrapper[4775]: I0321 05:02:21.645211 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dp2c\" (UniqueName: \"kubernetes.io/projected/65fae7ea-6284-4905-aa68-3251c5ef0602-kube-api-access-5dp2c\") pod \"65fae7ea-6284-4905-aa68-3251c5ef0602\" (UID: \"65fae7ea-6284-4905-aa68-3251c5ef0602\") " Mar 21 05:02:21 crc kubenswrapper[4775]: I0321 05:02:21.645294 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65fae7ea-6284-4905-aa68-3251c5ef0602-catalog-content\") pod \"65fae7ea-6284-4905-aa68-3251c5ef0602\" (UID: \"65fae7ea-6284-4905-aa68-3251c5ef0602\") " Mar 21 05:02:21 crc kubenswrapper[4775]: I0321 05:02:21.645328 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65fae7ea-6284-4905-aa68-3251c5ef0602-utilities\") pod \"65fae7ea-6284-4905-aa68-3251c5ef0602\" (UID: \"65fae7ea-6284-4905-aa68-3251c5ef0602\") " Mar 21 05:02:21 crc kubenswrapper[4775]: I0321 05:02:21.646283 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65fae7ea-6284-4905-aa68-3251c5ef0602-utilities" (OuterVolumeSpecName: "utilities") pod "65fae7ea-6284-4905-aa68-3251c5ef0602" (UID: "65fae7ea-6284-4905-aa68-3251c5ef0602"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:02:21 crc kubenswrapper[4775]: I0321 05:02:21.660529 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65fae7ea-6284-4905-aa68-3251c5ef0602-kube-api-access-5dp2c" (OuterVolumeSpecName: "kube-api-access-5dp2c") pod "65fae7ea-6284-4905-aa68-3251c5ef0602" (UID: "65fae7ea-6284-4905-aa68-3251c5ef0602"). InnerVolumeSpecName "kube-api-access-5dp2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:02:21 crc kubenswrapper[4775]: I0321 05:02:21.695680 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65fae7ea-6284-4905-aa68-3251c5ef0602-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "65fae7ea-6284-4905-aa68-3251c5ef0602" (UID: "65fae7ea-6284-4905-aa68-3251c5ef0602"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:02:21 crc kubenswrapper[4775]: I0321 05:02:21.747537 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dp2c\" (UniqueName: \"kubernetes.io/projected/65fae7ea-6284-4905-aa68-3251c5ef0602-kube-api-access-5dp2c\") on node \"crc\" DevicePath \"\"" Mar 21 05:02:21 crc kubenswrapper[4775]: I0321 05:02:21.747570 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65fae7ea-6284-4905-aa68-3251c5ef0602-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:02:21 crc kubenswrapper[4775]: I0321 05:02:21.747588 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65fae7ea-6284-4905-aa68-3251c5ef0602-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:02:22 crc kubenswrapper[4775]: I0321 05:02:22.048541 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hxvz8" event={"ID":"65fae7ea-6284-4905-aa68-3251c5ef0602","Type":"ContainerDied","Data":"b496a0cb17e25c9afc1e8fbd644c8d6a7c4698da16d39e00bebd370b5d3a0cc3"} Mar 21 05:02:22 crc kubenswrapper[4775]: I0321 05:02:22.048591 4775 scope.go:117] "RemoveContainer" containerID="9cec0f381408ed32105d216b30f0489513e66921ad73f2364aedcec396fd2f22" Mar 21 05:02:22 crc kubenswrapper[4775]: I0321 05:02:22.048604 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hxvz8" Mar 21 05:02:22 crc kubenswrapper[4775]: I0321 05:02:22.068156 4775 scope.go:117] "RemoveContainer" containerID="e6b7d92e6c0e684618dc387b9902e9548ae9e2cf09c7362448b678bb11ed2975" Mar 21 05:02:22 crc kubenswrapper[4775]: I0321 05:02:22.079700 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hxvz8"] Mar 21 05:02:22 crc kubenswrapper[4775]: I0321 05:02:22.086642 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hxvz8"] Mar 21 05:02:22 crc kubenswrapper[4775]: I0321 05:02:22.103649 4775 scope.go:117] "RemoveContainer" containerID="cc6575c3206e4931e90f2d2eeedfd610fe0c60e3f48437becf4f5af661bb2558" Mar 21 05:02:23 crc kubenswrapper[4775]: I0321 05:02:23.668973 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65fae7ea-6284-4905-aa68-3251c5ef0602" path="/var/lib/kubelet/pods/65fae7ea-6284-4905-aa68-3251c5ef0602/volumes" Mar 21 05:02:23 crc kubenswrapper[4775]: I0321 05:02:23.896799 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6695f56dbb-f6tqn" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.548902 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-hmmgn"] Mar 21 05:02:24 crc kubenswrapper[4775]: E0321 05:02:24.549201 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65fae7ea-6284-4905-aa68-3251c5ef0602" containerName="extract-utilities" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.549221 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="65fae7ea-6284-4905-aa68-3251c5ef0602" containerName="extract-utilities" Mar 21 05:02:24 crc kubenswrapper[4775]: E0321 05:02:24.549236 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65fae7ea-6284-4905-aa68-3251c5ef0602" containerName="registry-server" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.549245 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="65fae7ea-6284-4905-aa68-3251c5ef0602" containerName="registry-server" Mar 21 05:02:24 crc kubenswrapper[4775]: E0321 05:02:24.549267 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65fae7ea-6284-4905-aa68-3251c5ef0602" containerName="extract-content" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.549283 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="65fae7ea-6284-4905-aa68-3251c5ef0602" containerName="extract-content" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.549468 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="65fae7ea-6284-4905-aa68-3251c5ef0602" containerName="registry-server" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.552383 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.556013 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.556708 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-zh4bc" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.556730 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.564642 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-7jsnx"] Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.565754 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-7jsnx" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.567822 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.575022 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-7jsnx"] Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.586628 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9mrv\" (UniqueName: \"kubernetes.io/projected/e5808d9b-074a-4948-8283-fdfea77c63bc-kube-api-access-n9mrv\") pod \"frr-k8s-webhook-server-bcc4b6f68-7jsnx\" (UID: \"e5808d9b-074a-4948-8283-fdfea77c63bc\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-7jsnx" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.586700 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/530a34fb-bf82-4a2e-afdd-ec646afdebcd-frr-sockets\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.586721 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/530a34fb-bf82-4a2e-afdd-ec646afdebcd-frr-conf\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.586818 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/530a34fb-bf82-4a2e-afdd-ec646afdebcd-metrics\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.586850 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb66z\" (UniqueName: \"kubernetes.io/projected/530a34fb-bf82-4a2e-afdd-ec646afdebcd-kube-api-access-rb66z\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.586870 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e5808d9b-074a-4948-8283-fdfea77c63bc-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-7jsnx\" (UID: \"e5808d9b-074a-4948-8283-fdfea77c63bc\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-7jsnx" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.586986 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/530a34fb-bf82-4a2e-afdd-ec646afdebcd-metrics-certs\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.587016 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/530a34fb-bf82-4a2e-afdd-ec646afdebcd-frr-startup\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.587130 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/530a34fb-bf82-4a2e-afdd-ec646afdebcd-reloader\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.630885 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-cpw6m"] Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.631751 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-cpw6m" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.633129 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.633887 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.633909 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.634570 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-h6mzq" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.649061 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-7bb4cc7c98-qq976"] Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.650190 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-qq976" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.651748 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.674251 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-qq976"] Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.688472 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rwcj\" (UniqueName: \"kubernetes.io/projected/1d9349cc-e186-40c5-bb71-c176ff4f0a0d-kube-api-access-7rwcj\") pod \"speaker-cpw6m\" (UID: \"1d9349cc-e186-40c5-bb71-c176ff4f0a0d\") " pod="metallb-system/speaker-cpw6m" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.688520 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9mrv\" (UniqueName: \"kubernetes.io/projected/e5808d9b-074a-4948-8283-fdfea77c63bc-kube-api-access-n9mrv\") pod \"frr-k8s-webhook-server-bcc4b6f68-7jsnx\" (UID: \"e5808d9b-074a-4948-8283-fdfea77c63bc\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-7jsnx" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.688563 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/530a34fb-bf82-4a2e-afdd-ec646afdebcd-frr-sockets\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.688588 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/530a34fb-bf82-4a2e-afdd-ec646afdebcd-frr-conf\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.688628 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/530a34fb-bf82-4a2e-afdd-ec646afdebcd-metrics\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.688656 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb66z\" (UniqueName: \"kubernetes.io/projected/530a34fb-bf82-4a2e-afdd-ec646afdebcd-kube-api-access-rb66z\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.688682 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e5808d9b-074a-4948-8283-fdfea77c63bc-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-7jsnx\" (UID: \"e5808d9b-074a-4948-8283-fdfea77c63bc\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-7jsnx" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.688722 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1d9349cc-e186-40c5-bb71-c176ff4f0a0d-metrics-certs\") pod \"speaker-cpw6m\" (UID: \"1d9349cc-e186-40c5-bb71-c176ff4f0a0d\") " pod="metallb-system/speaker-cpw6m" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.688750 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1d9349cc-e186-40c5-bb71-c176ff4f0a0d-metallb-excludel2\") pod \"speaker-cpw6m\" (UID: \"1d9349cc-e186-40c5-bb71-c176ff4f0a0d\") " pod="metallb-system/speaker-cpw6m" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.688772 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjn5t\" (UniqueName: \"kubernetes.io/projected/ac0be1f3-95f0-40a4-9a94-c74cdaad9590-kube-api-access-gjn5t\") pod \"controller-7bb4cc7c98-qq976\" (UID: \"ac0be1f3-95f0-40a4-9a94-c74cdaad9590\") " pod="metallb-system/controller-7bb4cc7c98-qq976" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.688797 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ac0be1f3-95f0-40a4-9a94-c74cdaad9590-cert\") pod \"controller-7bb4cc7c98-qq976\" (UID: \"ac0be1f3-95f0-40a4-9a94-c74cdaad9590\") " pod="metallb-system/controller-7bb4cc7c98-qq976" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.688826 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1d9349cc-e186-40c5-bb71-c176ff4f0a0d-memberlist\") pod \"speaker-cpw6m\" (UID: \"1d9349cc-e186-40c5-bb71-c176ff4f0a0d\") " pod="metallb-system/speaker-cpw6m" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.688860 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ac0be1f3-95f0-40a4-9a94-c74cdaad9590-metrics-certs\") pod \"controller-7bb4cc7c98-qq976\" (UID: \"ac0be1f3-95f0-40a4-9a94-c74cdaad9590\") " pod="metallb-system/controller-7bb4cc7c98-qq976" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.688892 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/530a34fb-bf82-4a2e-afdd-ec646afdebcd-metrics-certs\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.688916 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/530a34fb-bf82-4a2e-afdd-ec646afdebcd-frr-startup\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.688967 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/530a34fb-bf82-4a2e-afdd-ec646afdebcd-reloader\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.688974 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/530a34fb-bf82-4a2e-afdd-ec646afdebcd-frr-conf\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: E0321 05:02:24.688985 4775 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.689021 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/530a34fb-bf82-4a2e-afdd-ec646afdebcd-metrics\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: E0321 05:02:24.689044 4775 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.689036 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/530a34fb-bf82-4a2e-afdd-ec646afdebcd-frr-sockets\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: E0321 05:02:24.689094 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/530a34fb-bf82-4a2e-afdd-ec646afdebcd-metrics-certs podName:530a34fb-bf82-4a2e-afdd-ec646afdebcd nodeName:}" failed. No retries permitted until 2026-03-21 05:02:25.189076926 +0000 UTC m=+898.165540540 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/530a34fb-bf82-4a2e-afdd-ec646afdebcd-metrics-certs") pod "frr-k8s-hmmgn" (UID: "530a34fb-bf82-4a2e-afdd-ec646afdebcd") : secret "frr-k8s-certs-secret" not found Mar 21 05:02:24 crc kubenswrapper[4775]: E0321 05:02:24.689129 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5808d9b-074a-4948-8283-fdfea77c63bc-cert podName:e5808d9b-074a-4948-8283-fdfea77c63bc nodeName:}" failed. No retries permitted until 2026-03-21 05:02:25.189105726 +0000 UTC m=+898.165569350 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e5808d9b-074a-4948-8283-fdfea77c63bc-cert") pod "frr-k8s-webhook-server-bcc4b6f68-7jsnx" (UID: "e5808d9b-074a-4948-8283-fdfea77c63bc") : secret "frr-k8s-webhook-server-cert" not found Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.689253 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/530a34fb-bf82-4a2e-afdd-ec646afdebcd-reloader\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.689883 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/530a34fb-bf82-4a2e-afdd-ec646afdebcd-frr-startup\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.712810 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9mrv\" (UniqueName: \"kubernetes.io/projected/e5808d9b-074a-4948-8283-fdfea77c63bc-kube-api-access-n9mrv\") pod \"frr-k8s-webhook-server-bcc4b6f68-7jsnx\" (UID: \"e5808d9b-074a-4948-8283-fdfea77c63bc\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-7jsnx" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.720659 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb66z\" (UniqueName: \"kubernetes.io/projected/530a34fb-bf82-4a2e-afdd-ec646afdebcd-kube-api-access-rb66z\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.789736 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1d9349cc-e186-40c5-bb71-c176ff4f0a0d-metallb-excludel2\") pod \"speaker-cpw6m\" (UID: \"1d9349cc-e186-40c5-bb71-c176ff4f0a0d\") " pod="metallb-system/speaker-cpw6m" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.789786 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjn5t\" (UniqueName: \"kubernetes.io/projected/ac0be1f3-95f0-40a4-9a94-c74cdaad9590-kube-api-access-gjn5t\") pod \"controller-7bb4cc7c98-qq976\" (UID: \"ac0be1f3-95f0-40a4-9a94-c74cdaad9590\") " pod="metallb-system/controller-7bb4cc7c98-qq976" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.789808 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ac0be1f3-95f0-40a4-9a94-c74cdaad9590-cert\") pod \"controller-7bb4cc7c98-qq976\" (UID: \"ac0be1f3-95f0-40a4-9a94-c74cdaad9590\") " pod="metallb-system/controller-7bb4cc7c98-qq976" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.789831 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1d9349cc-e186-40c5-bb71-c176ff4f0a0d-memberlist\") pod \"speaker-cpw6m\" (UID: \"1d9349cc-e186-40c5-bb71-c176ff4f0a0d\") " pod="metallb-system/speaker-cpw6m" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.789849 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ac0be1f3-95f0-40a4-9a94-c74cdaad9590-metrics-certs\") pod \"controller-7bb4cc7c98-qq976\" (UID: \"ac0be1f3-95f0-40a4-9a94-c74cdaad9590\") " pod="metallb-system/controller-7bb4cc7c98-qq976" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.789920 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rwcj\" (UniqueName: \"kubernetes.io/projected/1d9349cc-e186-40c5-bb71-c176ff4f0a0d-kube-api-access-7rwcj\") pod \"speaker-cpw6m\" (UID: \"1d9349cc-e186-40c5-bb71-c176ff4f0a0d\") " pod="metallb-system/speaker-cpw6m" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.789986 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1d9349cc-e186-40c5-bb71-c176ff4f0a0d-metrics-certs\") pod \"speaker-cpw6m\" (UID: \"1d9349cc-e186-40c5-bb71-c176ff4f0a0d\") " pod="metallb-system/speaker-cpw6m" Mar 21 05:02:24 crc kubenswrapper[4775]: E0321 05:02:24.790045 4775 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 21 05:02:24 crc kubenswrapper[4775]: E0321 05:02:24.790144 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1d9349cc-e186-40c5-bb71-c176ff4f0a0d-memberlist podName:1d9349cc-e186-40c5-bb71-c176ff4f0a0d nodeName:}" failed. No retries permitted until 2026-03-21 05:02:25.29010457 +0000 UTC m=+898.266568184 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1d9349cc-e186-40c5-bb71-c176ff4f0a0d-memberlist") pod "speaker-cpw6m" (UID: "1d9349cc-e186-40c5-bb71-c176ff4f0a0d") : secret "metallb-memberlist" not found Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.791020 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1d9349cc-e186-40c5-bb71-c176ff4f0a0d-metallb-excludel2\") pod \"speaker-cpw6m\" (UID: \"1d9349cc-e186-40c5-bb71-c176ff4f0a0d\") " pod="metallb-system/speaker-cpw6m" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.793235 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ac0be1f3-95f0-40a4-9a94-c74cdaad9590-metrics-certs\") pod \"controller-7bb4cc7c98-qq976\" (UID: \"ac0be1f3-95f0-40a4-9a94-c74cdaad9590\") " pod="metallb-system/controller-7bb4cc7c98-qq976" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.793677 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ac0be1f3-95f0-40a4-9a94-c74cdaad9590-cert\") pod \"controller-7bb4cc7c98-qq976\" (UID: \"ac0be1f3-95f0-40a4-9a94-c74cdaad9590\") " pod="metallb-system/controller-7bb4cc7c98-qq976" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.796586 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1d9349cc-e186-40c5-bb71-c176ff4f0a0d-metrics-certs\") pod \"speaker-cpw6m\" (UID: \"1d9349cc-e186-40c5-bb71-c176ff4f0a0d\") " pod="metallb-system/speaker-cpw6m" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.809581 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rwcj\" (UniqueName: \"kubernetes.io/projected/1d9349cc-e186-40c5-bb71-c176ff4f0a0d-kube-api-access-7rwcj\") pod \"speaker-cpw6m\" (UID: \"1d9349cc-e186-40c5-bb71-c176ff4f0a0d\") " pod="metallb-system/speaker-cpw6m" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.811923 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjn5t\" (UniqueName: \"kubernetes.io/projected/ac0be1f3-95f0-40a4-9a94-c74cdaad9590-kube-api-access-gjn5t\") pod \"controller-7bb4cc7c98-qq976\" (UID: \"ac0be1f3-95f0-40a4-9a94-c74cdaad9590\") " pod="metallb-system/controller-7bb4cc7c98-qq976" Mar 21 05:02:24 crc kubenswrapper[4775]: I0321 05:02:24.977894 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-qq976" Mar 21 05:02:25 crc kubenswrapper[4775]: I0321 05:02:25.195347 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/530a34fb-bf82-4a2e-afdd-ec646afdebcd-metrics-certs\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:25 crc kubenswrapper[4775]: I0321 05:02:25.196455 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e5808d9b-074a-4948-8283-fdfea77c63bc-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-7jsnx\" (UID: \"e5808d9b-074a-4948-8283-fdfea77c63bc\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-7jsnx" Mar 21 05:02:25 crc kubenswrapper[4775]: I0321 05:02:25.199846 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/530a34fb-bf82-4a2e-afdd-ec646afdebcd-metrics-certs\") pod \"frr-k8s-hmmgn\" (UID: \"530a34fb-bf82-4a2e-afdd-ec646afdebcd\") " pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:25 crc kubenswrapper[4775]: I0321 05:02:25.199944 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e5808d9b-074a-4948-8283-fdfea77c63bc-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-7jsnx\" (UID: \"e5808d9b-074a-4948-8283-fdfea77c63bc\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-7jsnx" Mar 21 05:02:25 crc kubenswrapper[4775]: I0321 05:02:25.297161 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1d9349cc-e186-40c5-bb71-c176ff4f0a0d-memberlist\") pod \"speaker-cpw6m\" (UID: \"1d9349cc-e186-40c5-bb71-c176ff4f0a0d\") " pod="metallb-system/speaker-cpw6m" Mar 21 05:02:25 crc kubenswrapper[4775]: E0321 05:02:25.297325 4775 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 21 05:02:25 crc kubenswrapper[4775]: E0321 05:02:25.297379 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1d9349cc-e186-40c5-bb71-c176ff4f0a0d-memberlist podName:1d9349cc-e186-40c5-bb71-c176ff4f0a0d nodeName:}" failed. No retries permitted until 2026-03-21 05:02:26.29736365 +0000 UTC m=+899.273827284 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1d9349cc-e186-40c5-bb71-c176ff4f0a0d-memberlist") pod "speaker-cpw6m" (UID: "1d9349cc-e186-40c5-bb71-c176ff4f0a0d") : secret "metallb-memberlist" not found Mar 21 05:02:25 crc kubenswrapper[4775]: I0321 05:02:25.412038 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-qq976"] Mar 21 05:02:25 crc kubenswrapper[4775]: I0321 05:02:25.480632 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:25 crc kubenswrapper[4775]: I0321 05:02:25.486551 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-7jsnx" Mar 21 05:02:25 crc kubenswrapper[4775]: I0321 05:02:25.744046 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-7jsnx"] Mar 21 05:02:25 crc kubenswrapper[4775]: W0321 05:02:25.758847 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5808d9b_074a_4948_8283_fdfea77c63bc.slice/crio-cb8df94439fc36f44eb6b9af7bbbaeee29cb06bd6c4d6c4f6cda37fa4cedab88 WatchSource:0}: Error finding container cb8df94439fc36f44eb6b9af7bbbaeee29cb06bd6c4d6c4f6cda37fa4cedab88: Status 404 returned error can't find the container with id cb8df94439fc36f44eb6b9af7bbbaeee29cb06bd6c4d6c4f6cda37fa4cedab88 Mar 21 05:02:26 crc kubenswrapper[4775]: I0321 05:02:26.072426 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-7jsnx" event={"ID":"e5808d9b-074a-4948-8283-fdfea77c63bc","Type":"ContainerStarted","Data":"cb8df94439fc36f44eb6b9af7bbbaeee29cb06bd6c4d6c4f6cda37fa4cedab88"} Mar 21 05:02:26 crc kubenswrapper[4775]: I0321 05:02:26.073752 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-qq976" event={"ID":"ac0be1f3-95f0-40a4-9a94-c74cdaad9590","Type":"ContainerStarted","Data":"a7d8aee5eb9ff6cafdeb4afc831b28835e283677be48c217812f06959791a63b"} Mar 21 05:02:26 crc kubenswrapper[4775]: I0321 05:02:26.073797 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-qq976" event={"ID":"ac0be1f3-95f0-40a4-9a94-c74cdaad9590","Type":"ContainerStarted","Data":"de8493125539d9d7f5ce1ea7208ef751013cd2467f41b90d3d349712234d4cd4"} Mar 21 05:02:26 crc kubenswrapper[4775]: I0321 05:02:26.075816 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hmmgn" event={"ID":"530a34fb-bf82-4a2e-afdd-ec646afdebcd","Type":"ContainerStarted","Data":"89337512b552921c50f81be5c9f601e1c5d7a75e332342a9762c46267d45f24b"} Mar 21 05:02:26 crc kubenswrapper[4775]: I0321 05:02:26.307963 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1d9349cc-e186-40c5-bb71-c176ff4f0a0d-memberlist\") pod \"speaker-cpw6m\" (UID: \"1d9349cc-e186-40c5-bb71-c176ff4f0a0d\") " pod="metallb-system/speaker-cpw6m" Mar 21 05:02:26 crc kubenswrapper[4775]: I0321 05:02:26.313540 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1d9349cc-e186-40c5-bb71-c176ff4f0a0d-memberlist\") pod \"speaker-cpw6m\" (UID: \"1d9349cc-e186-40c5-bb71-c176ff4f0a0d\") " pod="metallb-system/speaker-cpw6m" Mar 21 05:02:26 crc kubenswrapper[4775]: I0321 05:02:26.446575 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-cpw6m" Mar 21 05:02:26 crc kubenswrapper[4775]: W0321 05:02:26.479073 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d9349cc_e186_40c5_bb71_c176ff4f0a0d.slice/crio-94e56210d51a7672e1ae716a39d6eef54de6d76a15c7e07a6fff8c737a98462e WatchSource:0}: Error finding container 94e56210d51a7672e1ae716a39d6eef54de6d76a15c7e07a6fff8c737a98462e: Status 404 returned error can't find the container with id 94e56210d51a7672e1ae716a39d6eef54de6d76a15c7e07a6fff8c737a98462e Mar 21 05:02:27 crc kubenswrapper[4775]: I0321 05:02:27.082667 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-qq976" event={"ID":"ac0be1f3-95f0-40a4-9a94-c74cdaad9590","Type":"ContainerStarted","Data":"5cb3f47ea90c47d8cc0e49911dfa2b8add574ff8d691407a25c8af4fd435cd17"} Mar 21 05:02:27 crc kubenswrapper[4775]: I0321 05:02:27.083141 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-qq976" Mar 21 05:02:27 crc kubenswrapper[4775]: I0321 05:02:27.084518 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cpw6m" event={"ID":"1d9349cc-e186-40c5-bb71-c176ff4f0a0d","Type":"ContainerStarted","Data":"94e56210d51a7672e1ae716a39d6eef54de6d76a15c7e07a6fff8c737a98462e"} Mar 21 05:02:27 crc kubenswrapper[4775]: I0321 05:02:27.099519 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-7bb4cc7c98-qq976" podStartSLOduration=3.099504323 podStartE2EDuration="3.099504323s" podCreationTimestamp="2026-03-21 05:02:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:02:27.097757364 +0000 UTC m=+900.074220988" watchObservedRunningTime="2026-03-21 05:02:27.099504323 +0000 UTC m=+900.075967947" Mar 21 05:02:28 crc kubenswrapper[4775]: I0321 05:02:28.128774 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cpw6m" event={"ID":"1d9349cc-e186-40c5-bb71-c176ff4f0a0d","Type":"ContainerStarted","Data":"90c553183b0c8e95182d23edacc52df10433cabc4d0076042525c078b0df1f45"} Mar 21 05:02:28 crc kubenswrapper[4775]: I0321 05:02:28.129198 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cpw6m" event={"ID":"1d9349cc-e186-40c5-bb71-c176ff4f0a0d","Type":"ContainerStarted","Data":"a5c54ae33d6f574b7e40b16d5dd7c88eff1c53c7d889c60ba90b1df5580a637e"} Mar 21 05:02:28 crc kubenswrapper[4775]: I0321 05:02:28.153529 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-cpw6m" podStartSLOduration=4.153513779 podStartE2EDuration="4.153513779s" podCreationTimestamp="2026-03-21 05:02:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:02:28.150221046 +0000 UTC m=+901.126684690" watchObservedRunningTime="2026-03-21 05:02:28.153513779 +0000 UTC m=+901.129977403" Mar 21 05:02:29 crc kubenswrapper[4775]: I0321 05:02:29.145347 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-cpw6m" Mar 21 05:02:34 crc kubenswrapper[4775]: I0321 05:02:34.179469 4775 generic.go:334] "Generic (PLEG): container finished" podID="530a34fb-bf82-4a2e-afdd-ec646afdebcd" containerID="54e259d8125bddf7796cb32d7ca683f2f4184703473bd38e4b6a78ad9f9526e6" exitCode=0 Mar 21 05:02:34 crc kubenswrapper[4775]: I0321 05:02:34.179529 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hmmgn" event={"ID":"530a34fb-bf82-4a2e-afdd-ec646afdebcd","Type":"ContainerDied","Data":"54e259d8125bddf7796cb32d7ca683f2f4184703473bd38e4b6a78ad9f9526e6"} Mar 21 05:02:34 crc kubenswrapper[4775]: I0321 05:02:34.181916 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-7jsnx" event={"ID":"e5808d9b-074a-4948-8283-fdfea77c63bc","Type":"ContainerStarted","Data":"8f1a6980a0fec05028213e7a314e298e7e837a36ae032942202c882dc6f3a3b3"} Mar 21 05:02:34 crc kubenswrapper[4775]: I0321 05:02:34.182325 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-7jsnx" Mar 21 05:02:34 crc kubenswrapper[4775]: I0321 05:02:34.225004 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-7jsnx" podStartSLOduration=2.867722845 podStartE2EDuration="10.224983585s" podCreationTimestamp="2026-03-21 05:02:24 +0000 UTC" firstStartedPulling="2026-03-21 05:02:25.76564171 +0000 UTC m=+898.742105334" lastFinishedPulling="2026-03-21 05:02:33.12290245 +0000 UTC m=+906.099366074" observedRunningTime="2026-03-21 05:02:34.215880548 +0000 UTC m=+907.192344162" watchObservedRunningTime="2026-03-21 05:02:34.224983585 +0000 UTC m=+907.201447209" Mar 21 05:02:35 crc kubenswrapper[4775]: I0321 05:02:35.189868 4775 generic.go:334] "Generic (PLEG): container finished" podID="530a34fb-bf82-4a2e-afdd-ec646afdebcd" containerID="145d9a2ceda338b7345a9b0fa7c873ac6f653688019db2c535c7a4c272788f3b" exitCode=0 Mar 21 05:02:35 crc kubenswrapper[4775]: I0321 05:02:35.189967 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hmmgn" event={"ID":"530a34fb-bf82-4a2e-afdd-ec646afdebcd","Type":"ContainerDied","Data":"145d9a2ceda338b7345a9b0fa7c873ac6f653688019db2c535c7a4c272788f3b"} Mar 21 05:02:36 crc kubenswrapper[4775]: I0321 05:02:36.197907 4775 generic.go:334] "Generic (PLEG): container finished" podID="530a34fb-bf82-4a2e-afdd-ec646afdebcd" containerID="0c54e8f5c1e6b7dea5786eaa31e7e8687b528834571742012096886361314ce6" exitCode=0 Mar 21 05:02:36 crc kubenswrapper[4775]: I0321 05:02:36.197959 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hmmgn" event={"ID":"530a34fb-bf82-4a2e-afdd-ec646afdebcd","Type":"ContainerDied","Data":"0c54e8f5c1e6b7dea5786eaa31e7e8687b528834571742012096886361314ce6"} Mar 21 05:02:37 crc kubenswrapper[4775]: I0321 05:02:37.208816 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hmmgn" event={"ID":"530a34fb-bf82-4a2e-afdd-ec646afdebcd","Type":"ContainerStarted","Data":"9c42a35f97ea0ba130660183ffba8f33c9749b8aebc6913493caf59c6dc4cea6"} Mar 21 05:02:37 crc kubenswrapper[4775]: I0321 05:02:37.208862 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hmmgn" event={"ID":"530a34fb-bf82-4a2e-afdd-ec646afdebcd","Type":"ContainerStarted","Data":"9a626c5ffdaa4b0c5c78534d88a49bfc2781bd49f4ff6c2a0cf10147c501a196"} Mar 21 05:02:37 crc kubenswrapper[4775]: I0321 05:02:37.208871 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hmmgn" event={"ID":"530a34fb-bf82-4a2e-afdd-ec646afdebcd","Type":"ContainerStarted","Data":"fe0785423bed047ecc10f06350efd59ec25ba622f5842aebec09a977f496e536"} Mar 21 05:02:37 crc kubenswrapper[4775]: I0321 05:02:37.208882 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hmmgn" event={"ID":"530a34fb-bf82-4a2e-afdd-ec646afdebcd","Type":"ContainerStarted","Data":"d0d752395ccc084e37cf5240f02979af8e6401976260d9b82ba569336db0be16"} Mar 21 05:02:37 crc kubenswrapper[4775]: I0321 05:02:37.296978 4775 scope.go:117] "RemoveContainer" containerID="67295240b136b0bcf5b3e83b034399ed927b48c5da3ccf9341e8f2fa7f873c3a" Mar 21 05:02:38 crc kubenswrapper[4775]: I0321 05:02:38.219540 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hmmgn" event={"ID":"530a34fb-bf82-4a2e-afdd-ec646afdebcd","Type":"ContainerStarted","Data":"e53eb5608fb39d95b16699dfb80d0b50647087d0f2a635ee454e89cb381dea5a"} Mar 21 05:02:38 crc kubenswrapper[4775]: I0321 05:02:38.219801 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hmmgn" event={"ID":"530a34fb-bf82-4a2e-afdd-ec646afdebcd","Type":"ContainerStarted","Data":"b8d1c6e8deb07620cfda015d5497bf040582fad0feac2d48c532974f3c04be9b"} Mar 21 05:02:38 crc kubenswrapper[4775]: I0321 05:02:38.219838 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:38 crc kubenswrapper[4775]: I0321 05:02:38.247380 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-hmmgn" podStartSLOduration=6.968295908 podStartE2EDuration="14.247360679s" podCreationTimestamp="2026-03-21 05:02:24 +0000 UTC" firstStartedPulling="2026-03-21 05:02:25.820835399 +0000 UTC m=+898.797299023" lastFinishedPulling="2026-03-21 05:02:33.09990017 +0000 UTC m=+906.076363794" observedRunningTime="2026-03-21 05:02:38.244473227 +0000 UTC m=+911.220936861" watchObservedRunningTime="2026-03-21 05:02:38.247360679 +0000 UTC m=+911.223824323" Mar 21 05:02:40 crc kubenswrapper[4775]: I0321 05:02:40.481085 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:40 crc kubenswrapper[4775]: I0321 05:02:40.539745 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:44 crc kubenswrapper[4775]: I0321 05:02:44.996972 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-7bb4cc7c98-qq976" Mar 21 05:02:45 crc kubenswrapper[4775]: I0321 05:02:45.492555 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-7jsnx" Mar 21 05:02:46 crc kubenswrapper[4775]: I0321 05:02:46.450238 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-cpw6m" Mar 21 05:02:52 crc kubenswrapper[4775]: I0321 05:02:52.362666 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-7tr45"] Mar 21 05:02:52 crc kubenswrapper[4775]: I0321 05:02:52.364162 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7tr45" Mar 21 05:02:52 crc kubenswrapper[4775]: I0321 05:02:52.367504 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 21 05:02:52 crc kubenswrapper[4775]: I0321 05:02:52.369283 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 21 05:02:52 crc kubenswrapper[4775]: I0321 05:02:52.369541 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-4vkm5" Mar 21 05:02:52 crc kubenswrapper[4775]: I0321 05:02:52.373934 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7tr45"] Mar 21 05:02:52 crc kubenswrapper[4775]: I0321 05:02:52.446022 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42z9n\" (UniqueName: \"kubernetes.io/projected/c55e2ca9-69fd-41be-a5d8-838cdf259c0d-kube-api-access-42z9n\") pod \"openstack-operator-index-7tr45\" (UID: \"c55e2ca9-69fd-41be-a5d8-838cdf259c0d\") " pod="openstack-operators/openstack-operator-index-7tr45" Mar 21 05:02:52 crc kubenswrapper[4775]: I0321 05:02:52.547340 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42z9n\" (UniqueName: \"kubernetes.io/projected/c55e2ca9-69fd-41be-a5d8-838cdf259c0d-kube-api-access-42z9n\") pod \"openstack-operator-index-7tr45\" (UID: \"c55e2ca9-69fd-41be-a5d8-838cdf259c0d\") " pod="openstack-operators/openstack-operator-index-7tr45" Mar 21 05:02:52 crc kubenswrapper[4775]: I0321 05:02:52.575828 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42z9n\" (UniqueName: \"kubernetes.io/projected/c55e2ca9-69fd-41be-a5d8-838cdf259c0d-kube-api-access-42z9n\") pod \"openstack-operator-index-7tr45\" (UID: \"c55e2ca9-69fd-41be-a5d8-838cdf259c0d\") " pod="openstack-operators/openstack-operator-index-7tr45" Mar 21 05:02:52 crc kubenswrapper[4775]: I0321 05:02:52.682093 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7tr45" Mar 21 05:02:53 crc kubenswrapper[4775]: I0321 05:02:53.092862 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7tr45"] Mar 21 05:02:53 crc kubenswrapper[4775]: I0321 05:02:53.309740 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7tr45" event={"ID":"c55e2ca9-69fd-41be-a5d8-838cdf259c0d","Type":"ContainerStarted","Data":"995a3cedd5dcd750f59c7675e4bd7b49d6165f4247ebdf791eccf8293e7782e5"} Mar 21 05:02:55 crc kubenswrapper[4775]: I0321 05:02:55.483726 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-hmmgn" Mar 21 05:02:57 crc kubenswrapper[4775]: I0321 05:02:57.340650 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7tr45" event={"ID":"c55e2ca9-69fd-41be-a5d8-838cdf259c0d","Type":"ContainerStarted","Data":"685d7998a05afd3596f950c41f49e4d6db42d13a64fd7fd1656134c40c8626ab"} Mar 21 05:02:57 crc kubenswrapper[4775]: I0321 05:02:57.362968 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-7tr45" podStartSLOduration=2.163409864 podStartE2EDuration="5.362949941s" podCreationTimestamp="2026-03-21 05:02:52 +0000 UTC" firstStartedPulling="2026-03-21 05:02:53.109835956 +0000 UTC m=+926.086299590" lastFinishedPulling="2026-03-21 05:02:56.309376043 +0000 UTC m=+929.285839667" observedRunningTime="2026-03-21 05:02:57.360407079 +0000 UTC m=+930.336870703" watchObservedRunningTime="2026-03-21 05:02:57.362949941 +0000 UTC m=+930.339413575" Mar 21 05:02:57 crc kubenswrapper[4775]: I0321 05:02:57.749279 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-7tr45"] Mar 21 05:02:58 crc kubenswrapper[4775]: I0321 05:02:58.359161 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-hvtkg"] Mar 21 05:02:58 crc kubenswrapper[4775]: I0321 05:02:58.360023 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hvtkg" Mar 21 05:02:58 crc kubenswrapper[4775]: I0321 05:02:58.376417 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hvtkg"] Mar 21 05:02:58 crc kubenswrapper[4775]: I0321 05:02:58.424686 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn6hd\" (UniqueName: \"kubernetes.io/projected/8c7426e8-8cec-4c84-8810-03a091d87cd9-kube-api-access-kn6hd\") pod \"openstack-operator-index-hvtkg\" (UID: \"8c7426e8-8cec-4c84-8810-03a091d87cd9\") " pod="openstack-operators/openstack-operator-index-hvtkg" Mar 21 05:02:58 crc kubenswrapper[4775]: I0321 05:02:58.526452 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn6hd\" (UniqueName: \"kubernetes.io/projected/8c7426e8-8cec-4c84-8810-03a091d87cd9-kube-api-access-kn6hd\") pod \"openstack-operator-index-hvtkg\" (UID: \"8c7426e8-8cec-4c84-8810-03a091d87cd9\") " pod="openstack-operators/openstack-operator-index-hvtkg" Mar 21 05:02:58 crc kubenswrapper[4775]: I0321 05:02:58.545871 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn6hd\" (UniqueName: \"kubernetes.io/projected/8c7426e8-8cec-4c84-8810-03a091d87cd9-kube-api-access-kn6hd\") pod \"openstack-operator-index-hvtkg\" (UID: \"8c7426e8-8cec-4c84-8810-03a091d87cd9\") " pod="openstack-operators/openstack-operator-index-hvtkg" Mar 21 05:02:58 crc kubenswrapper[4775]: I0321 05:02:58.677636 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hvtkg" Mar 21 05:02:59 crc kubenswrapper[4775]: I0321 05:02:59.069680 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hvtkg"] Mar 21 05:02:59 crc kubenswrapper[4775]: I0321 05:02:59.354262 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-7tr45" podUID="c55e2ca9-69fd-41be-a5d8-838cdf259c0d" containerName="registry-server" containerID="cri-o://685d7998a05afd3596f950c41f49e4d6db42d13a64fd7fd1656134c40c8626ab" gracePeriod=2 Mar 21 05:02:59 crc kubenswrapper[4775]: I0321 05:02:59.354823 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hvtkg" event={"ID":"8c7426e8-8cec-4c84-8810-03a091d87cd9","Type":"ContainerStarted","Data":"5d31662d351f54fed982495f8f8148bc6d813e97a6151cf3f8cb22892e4542ad"} Mar 21 05:02:59 crc kubenswrapper[4775]: I0321 05:02:59.354845 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hvtkg" event={"ID":"8c7426e8-8cec-4c84-8810-03a091d87cd9","Type":"ContainerStarted","Data":"7a48142b15e92f4c26c5e5751402fda8d00b456235bff782d4510e971251e495"} Mar 21 05:02:59 crc kubenswrapper[4775]: I0321 05:02:59.382932 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-hvtkg" podStartSLOduration=1.344728895 podStartE2EDuration="1.382908067s" podCreationTimestamp="2026-03-21 05:02:58 +0000 UTC" firstStartedPulling="2026-03-21 05:02:59.067788476 +0000 UTC m=+932.044252100" lastFinishedPulling="2026-03-21 05:02:59.105967648 +0000 UTC m=+932.082431272" observedRunningTime="2026-03-21 05:02:59.377141744 +0000 UTC m=+932.353605358" watchObservedRunningTime="2026-03-21 05:02:59.382908067 +0000 UTC m=+932.359371701" Mar 21 05:02:59 crc kubenswrapper[4775]: I0321 05:02:59.726627 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7tr45" Mar 21 05:02:59 crc kubenswrapper[4775]: I0321 05:02:59.844715 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42z9n\" (UniqueName: \"kubernetes.io/projected/c55e2ca9-69fd-41be-a5d8-838cdf259c0d-kube-api-access-42z9n\") pod \"c55e2ca9-69fd-41be-a5d8-838cdf259c0d\" (UID: \"c55e2ca9-69fd-41be-a5d8-838cdf259c0d\") " Mar 21 05:02:59 crc kubenswrapper[4775]: I0321 05:02:59.851362 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c55e2ca9-69fd-41be-a5d8-838cdf259c0d-kube-api-access-42z9n" (OuterVolumeSpecName: "kube-api-access-42z9n") pod "c55e2ca9-69fd-41be-a5d8-838cdf259c0d" (UID: "c55e2ca9-69fd-41be-a5d8-838cdf259c0d"). InnerVolumeSpecName "kube-api-access-42z9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:02:59 crc kubenswrapper[4775]: I0321 05:02:59.946880 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42z9n\" (UniqueName: \"kubernetes.io/projected/c55e2ca9-69fd-41be-a5d8-838cdf259c0d-kube-api-access-42z9n\") on node \"crc\" DevicePath \"\"" Mar 21 05:03:00 crc kubenswrapper[4775]: I0321 05:03:00.364208 4775 generic.go:334] "Generic (PLEG): container finished" podID="c55e2ca9-69fd-41be-a5d8-838cdf259c0d" containerID="685d7998a05afd3596f950c41f49e4d6db42d13a64fd7fd1656134c40c8626ab" exitCode=0 Mar 21 05:03:00 crc kubenswrapper[4775]: I0321 05:03:00.364275 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7tr45" event={"ID":"c55e2ca9-69fd-41be-a5d8-838cdf259c0d","Type":"ContainerDied","Data":"685d7998a05afd3596f950c41f49e4d6db42d13a64fd7fd1656134c40c8626ab"} Mar 21 05:03:00 crc kubenswrapper[4775]: I0321 05:03:00.364329 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7tr45" Mar 21 05:03:00 crc kubenswrapper[4775]: I0321 05:03:00.364364 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7tr45" event={"ID":"c55e2ca9-69fd-41be-a5d8-838cdf259c0d","Type":"ContainerDied","Data":"995a3cedd5dcd750f59c7675e4bd7b49d6165f4247ebdf791eccf8293e7782e5"} Mar 21 05:03:00 crc kubenswrapper[4775]: I0321 05:03:00.364400 4775 scope.go:117] "RemoveContainer" containerID="685d7998a05afd3596f950c41f49e4d6db42d13a64fd7fd1656134c40c8626ab" Mar 21 05:03:00 crc kubenswrapper[4775]: I0321 05:03:00.390233 4775 scope.go:117] "RemoveContainer" containerID="685d7998a05afd3596f950c41f49e4d6db42d13a64fd7fd1656134c40c8626ab" Mar 21 05:03:00 crc kubenswrapper[4775]: E0321 05:03:00.391438 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"685d7998a05afd3596f950c41f49e4d6db42d13a64fd7fd1656134c40c8626ab\": container with ID starting with 685d7998a05afd3596f950c41f49e4d6db42d13a64fd7fd1656134c40c8626ab not found: ID does not exist" containerID="685d7998a05afd3596f950c41f49e4d6db42d13a64fd7fd1656134c40c8626ab" Mar 21 05:03:00 crc kubenswrapper[4775]: I0321 05:03:00.391505 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"685d7998a05afd3596f950c41f49e4d6db42d13a64fd7fd1656134c40c8626ab"} err="failed to get container status \"685d7998a05afd3596f950c41f49e4d6db42d13a64fd7fd1656134c40c8626ab\": rpc error: code = NotFound desc = could not find container \"685d7998a05afd3596f950c41f49e4d6db42d13a64fd7fd1656134c40c8626ab\": container with ID starting with 685d7998a05afd3596f950c41f49e4d6db42d13a64fd7fd1656134c40c8626ab not found: ID does not exist" Mar 21 05:03:00 crc kubenswrapper[4775]: I0321 05:03:00.405683 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-7tr45"] Mar 21 05:03:00 crc kubenswrapper[4775]: I0321 05:03:00.414742 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-7tr45"] Mar 21 05:03:01 crc kubenswrapper[4775]: I0321 05:03:01.675745 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c55e2ca9-69fd-41be-a5d8-838cdf259c0d" path="/var/lib/kubelet/pods/c55e2ca9-69fd-41be-a5d8-838cdf259c0d/volumes" Mar 21 05:03:08 crc kubenswrapper[4775]: I0321 05:03:08.678379 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-hvtkg" Mar 21 05:03:08 crc kubenswrapper[4775]: I0321 05:03:08.679494 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-hvtkg" Mar 21 05:03:08 crc kubenswrapper[4775]: I0321 05:03:08.713402 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-hvtkg" Mar 21 05:03:09 crc kubenswrapper[4775]: I0321 05:03:09.440434 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-hvtkg" Mar 21 05:03:15 crc kubenswrapper[4775]: I0321 05:03:15.701301 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt"] Mar 21 05:03:15 crc kubenswrapper[4775]: E0321 05:03:15.702030 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c55e2ca9-69fd-41be-a5d8-838cdf259c0d" containerName="registry-server" Mar 21 05:03:15 crc kubenswrapper[4775]: I0321 05:03:15.702041 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c55e2ca9-69fd-41be-a5d8-838cdf259c0d" containerName="registry-server" Mar 21 05:03:15 crc kubenswrapper[4775]: I0321 05:03:15.702171 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c55e2ca9-69fd-41be-a5d8-838cdf259c0d" containerName="registry-server" Mar 21 05:03:15 crc kubenswrapper[4775]: I0321 05:03:15.702993 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt" Mar 21 05:03:15 crc kubenswrapper[4775]: I0321 05:03:15.704904 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-8lvrj" Mar 21 05:03:15 crc kubenswrapper[4775]: I0321 05:03:15.708918 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt"] Mar 21 05:03:15 crc kubenswrapper[4775]: I0321 05:03:15.760096 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jd54\" (UniqueName: \"kubernetes.io/projected/a8157e0e-bf83-4dbd-af86-d09d14e6e1b1-kube-api-access-7jd54\") pod \"68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt\" (UID: \"a8157e0e-bf83-4dbd-af86-d09d14e6e1b1\") " pod="openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt" Mar 21 05:03:15 crc kubenswrapper[4775]: I0321 05:03:15.760163 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a8157e0e-bf83-4dbd-af86-d09d14e6e1b1-util\") pod \"68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt\" (UID: \"a8157e0e-bf83-4dbd-af86-d09d14e6e1b1\") " pod="openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt" Mar 21 05:03:15 crc kubenswrapper[4775]: I0321 05:03:15.760211 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a8157e0e-bf83-4dbd-af86-d09d14e6e1b1-bundle\") pod \"68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt\" (UID: \"a8157e0e-bf83-4dbd-af86-d09d14e6e1b1\") " pod="openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt" Mar 21 05:03:15 crc kubenswrapper[4775]: I0321 05:03:15.861611 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jd54\" (UniqueName: \"kubernetes.io/projected/a8157e0e-bf83-4dbd-af86-d09d14e6e1b1-kube-api-access-7jd54\") pod \"68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt\" (UID: \"a8157e0e-bf83-4dbd-af86-d09d14e6e1b1\") " pod="openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt" Mar 21 05:03:15 crc kubenswrapper[4775]: I0321 05:03:15.861661 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a8157e0e-bf83-4dbd-af86-d09d14e6e1b1-util\") pod \"68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt\" (UID: \"a8157e0e-bf83-4dbd-af86-d09d14e6e1b1\") " pod="openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt" Mar 21 05:03:15 crc kubenswrapper[4775]: I0321 05:03:15.861695 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a8157e0e-bf83-4dbd-af86-d09d14e6e1b1-bundle\") pod \"68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt\" (UID: \"a8157e0e-bf83-4dbd-af86-d09d14e6e1b1\") " pod="openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt" Mar 21 05:03:15 crc kubenswrapper[4775]: I0321 05:03:15.862218 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a8157e0e-bf83-4dbd-af86-d09d14e6e1b1-bundle\") pod \"68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt\" (UID: \"a8157e0e-bf83-4dbd-af86-d09d14e6e1b1\") " pod="openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt" Mar 21 05:03:15 crc kubenswrapper[4775]: I0321 05:03:15.862251 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a8157e0e-bf83-4dbd-af86-d09d14e6e1b1-util\") pod \"68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt\" (UID: \"a8157e0e-bf83-4dbd-af86-d09d14e6e1b1\") " pod="openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt" Mar 21 05:03:15 crc kubenswrapper[4775]: I0321 05:03:15.881306 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jd54\" (UniqueName: \"kubernetes.io/projected/a8157e0e-bf83-4dbd-af86-d09d14e6e1b1-kube-api-access-7jd54\") pod \"68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt\" (UID: \"a8157e0e-bf83-4dbd-af86-d09d14e6e1b1\") " pod="openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt" Mar 21 05:03:16 crc kubenswrapper[4775]: I0321 05:03:16.021238 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt" Mar 21 05:03:16 crc kubenswrapper[4775]: I0321 05:03:16.441806 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt"] Mar 21 05:03:16 crc kubenswrapper[4775]: I0321 05:03:16.460513 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt" event={"ID":"a8157e0e-bf83-4dbd-af86-d09d14e6e1b1","Type":"ContainerStarted","Data":"70feef91658f1da99ff064bdbcec3d2427164483c135492b6403332167eba6c0"} Mar 21 05:03:17 crc kubenswrapper[4775]: I0321 05:03:17.468607 4775 generic.go:334] "Generic (PLEG): container finished" podID="a8157e0e-bf83-4dbd-af86-d09d14e6e1b1" containerID="817da2dd74a7e5323813c63e67bc39fdde83314b3f9d77ceb23de398e14345ed" exitCode=0 Mar 21 05:03:17 crc kubenswrapper[4775]: I0321 05:03:17.468694 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt" event={"ID":"a8157e0e-bf83-4dbd-af86-d09d14e6e1b1","Type":"ContainerDied","Data":"817da2dd74a7e5323813c63e67bc39fdde83314b3f9d77ceb23de398e14345ed"} Mar 21 05:03:17 crc kubenswrapper[4775]: I0321 05:03:17.470950 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 05:03:18 crc kubenswrapper[4775]: I0321 05:03:18.477507 4775 generic.go:334] "Generic (PLEG): container finished" podID="a8157e0e-bf83-4dbd-af86-d09d14e6e1b1" containerID="2499d2237f0079f5f8fec2d33b699971213e7f38d15de3fca71a9a5053917b17" exitCode=0 Mar 21 05:03:18 crc kubenswrapper[4775]: I0321 05:03:18.478043 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt" event={"ID":"a8157e0e-bf83-4dbd-af86-d09d14e6e1b1","Type":"ContainerDied","Data":"2499d2237f0079f5f8fec2d33b699971213e7f38d15de3fca71a9a5053917b17"} Mar 21 05:03:19 crc kubenswrapper[4775]: I0321 05:03:19.487415 4775 generic.go:334] "Generic (PLEG): container finished" podID="a8157e0e-bf83-4dbd-af86-d09d14e6e1b1" containerID="22f4582caaafb23bc51a920330711510038ca7554591e01661392f782fa03521" exitCode=0 Mar 21 05:03:19 crc kubenswrapper[4775]: I0321 05:03:19.487463 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt" event={"ID":"a8157e0e-bf83-4dbd-af86-d09d14e6e1b1","Type":"ContainerDied","Data":"22f4582caaafb23bc51a920330711510038ca7554591e01661392f782fa03521"} Mar 21 05:03:20 crc kubenswrapper[4775]: I0321 05:03:20.763900 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt" Mar 21 05:03:20 crc kubenswrapper[4775]: I0321 05:03:20.862961 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a8157e0e-bf83-4dbd-af86-d09d14e6e1b1-bundle\") pod \"a8157e0e-bf83-4dbd-af86-d09d14e6e1b1\" (UID: \"a8157e0e-bf83-4dbd-af86-d09d14e6e1b1\") " Mar 21 05:03:20 crc kubenswrapper[4775]: I0321 05:03:20.863374 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jd54\" (UniqueName: \"kubernetes.io/projected/a8157e0e-bf83-4dbd-af86-d09d14e6e1b1-kube-api-access-7jd54\") pod \"a8157e0e-bf83-4dbd-af86-d09d14e6e1b1\" (UID: \"a8157e0e-bf83-4dbd-af86-d09d14e6e1b1\") " Mar 21 05:03:20 crc kubenswrapper[4775]: I0321 05:03:20.863404 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a8157e0e-bf83-4dbd-af86-d09d14e6e1b1-util\") pod \"a8157e0e-bf83-4dbd-af86-d09d14e6e1b1\" (UID: \"a8157e0e-bf83-4dbd-af86-d09d14e6e1b1\") " Mar 21 05:03:20 crc kubenswrapper[4775]: I0321 05:03:20.864355 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8157e0e-bf83-4dbd-af86-d09d14e6e1b1-bundle" (OuterVolumeSpecName: "bundle") pod "a8157e0e-bf83-4dbd-af86-d09d14e6e1b1" (UID: "a8157e0e-bf83-4dbd-af86-d09d14e6e1b1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:03:20 crc kubenswrapper[4775]: I0321 05:03:20.882925 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8157e0e-bf83-4dbd-af86-d09d14e6e1b1-util" (OuterVolumeSpecName: "util") pod "a8157e0e-bf83-4dbd-af86-d09d14e6e1b1" (UID: "a8157e0e-bf83-4dbd-af86-d09d14e6e1b1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:03:20 crc kubenswrapper[4775]: I0321 05:03:20.883340 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8157e0e-bf83-4dbd-af86-d09d14e6e1b1-kube-api-access-7jd54" (OuterVolumeSpecName: "kube-api-access-7jd54") pod "a8157e0e-bf83-4dbd-af86-d09d14e6e1b1" (UID: "a8157e0e-bf83-4dbd-af86-d09d14e6e1b1"). InnerVolumeSpecName "kube-api-access-7jd54". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:03:20 crc kubenswrapper[4775]: I0321 05:03:20.964573 4775 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a8157e0e-bf83-4dbd-af86-d09d14e6e1b1-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:03:20 crc kubenswrapper[4775]: I0321 05:03:20.964606 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jd54\" (UniqueName: \"kubernetes.io/projected/a8157e0e-bf83-4dbd-af86-d09d14e6e1b1-kube-api-access-7jd54\") on node \"crc\" DevicePath \"\"" Mar 21 05:03:20 crc kubenswrapper[4775]: I0321 05:03:20.964616 4775 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a8157e0e-bf83-4dbd-af86-d09d14e6e1b1-util\") on node \"crc\" DevicePath \"\"" Mar 21 05:03:21 crc kubenswrapper[4775]: I0321 05:03:21.502861 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt" event={"ID":"a8157e0e-bf83-4dbd-af86-d09d14e6e1b1","Type":"ContainerDied","Data":"70feef91658f1da99ff064bdbcec3d2427164483c135492b6403332167eba6c0"} Mar 21 05:03:21 crc kubenswrapper[4775]: I0321 05:03:21.502903 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70feef91658f1da99ff064bdbcec3d2427164483c135492b6403332167eba6c0" Mar 21 05:03:21 crc kubenswrapper[4775]: I0321 05:03:21.503012 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt" Mar 21 05:03:28 crc kubenswrapper[4775]: I0321 05:03:28.413965 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-85fcfb8fbb-q2k4k"] Mar 21 05:03:28 crc kubenswrapper[4775]: E0321 05:03:28.414975 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8157e0e-bf83-4dbd-af86-d09d14e6e1b1" containerName="pull" Mar 21 05:03:28 crc kubenswrapper[4775]: I0321 05:03:28.414998 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8157e0e-bf83-4dbd-af86-d09d14e6e1b1" containerName="pull" Mar 21 05:03:28 crc kubenswrapper[4775]: E0321 05:03:28.415019 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8157e0e-bf83-4dbd-af86-d09d14e6e1b1" containerName="extract" Mar 21 05:03:28 crc kubenswrapper[4775]: I0321 05:03:28.415030 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8157e0e-bf83-4dbd-af86-d09d14e6e1b1" containerName="extract" Mar 21 05:03:28 crc kubenswrapper[4775]: E0321 05:03:28.415047 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8157e0e-bf83-4dbd-af86-d09d14e6e1b1" containerName="util" Mar 21 05:03:28 crc kubenswrapper[4775]: I0321 05:03:28.415060 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8157e0e-bf83-4dbd-af86-d09d14e6e1b1" containerName="util" Mar 21 05:03:28 crc kubenswrapper[4775]: I0321 05:03:28.415287 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8157e0e-bf83-4dbd-af86-d09d14e6e1b1" containerName="extract" Mar 21 05:03:28 crc kubenswrapper[4775]: I0321 05:03:28.415915 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-85fcfb8fbb-q2k4k" Mar 21 05:03:28 crc kubenswrapper[4775]: I0321 05:03:28.418420 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-2cms9" Mar 21 05:03:28 crc kubenswrapper[4775]: I0321 05:03:28.456923 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-85fcfb8fbb-q2k4k"] Mar 21 05:03:28 crc kubenswrapper[4775]: I0321 05:03:28.568670 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9shlv\" (UniqueName: \"kubernetes.io/projected/899f7d20-7208-419e-b0f8-36c7fbf2e841-kube-api-access-9shlv\") pod \"openstack-operator-controller-init-85fcfb8fbb-q2k4k\" (UID: \"899f7d20-7208-419e-b0f8-36c7fbf2e841\") " pod="openstack-operators/openstack-operator-controller-init-85fcfb8fbb-q2k4k" Mar 21 05:03:28 crc kubenswrapper[4775]: I0321 05:03:28.669692 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9shlv\" (UniqueName: \"kubernetes.io/projected/899f7d20-7208-419e-b0f8-36c7fbf2e841-kube-api-access-9shlv\") pod \"openstack-operator-controller-init-85fcfb8fbb-q2k4k\" (UID: \"899f7d20-7208-419e-b0f8-36c7fbf2e841\") " pod="openstack-operators/openstack-operator-controller-init-85fcfb8fbb-q2k4k" Mar 21 05:03:28 crc kubenswrapper[4775]: I0321 05:03:28.698452 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9shlv\" (UniqueName: \"kubernetes.io/projected/899f7d20-7208-419e-b0f8-36c7fbf2e841-kube-api-access-9shlv\") pod \"openstack-operator-controller-init-85fcfb8fbb-q2k4k\" (UID: \"899f7d20-7208-419e-b0f8-36c7fbf2e841\") " pod="openstack-operators/openstack-operator-controller-init-85fcfb8fbb-q2k4k" Mar 21 05:03:28 crc kubenswrapper[4775]: I0321 05:03:28.737059 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-85fcfb8fbb-q2k4k" Mar 21 05:03:29 crc kubenswrapper[4775]: I0321 05:03:29.166655 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-85fcfb8fbb-q2k4k"] Mar 21 05:03:29 crc kubenswrapper[4775]: I0321 05:03:29.553159 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-85fcfb8fbb-q2k4k" event={"ID":"899f7d20-7208-419e-b0f8-36c7fbf2e841","Type":"ContainerStarted","Data":"e8d898aaf010ff6001c1f8bad0dc77465c515e5a320cf98c275b33204e0f44d0"} Mar 21 05:03:32 crc kubenswrapper[4775]: I0321 05:03:32.482821 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:03:32 crc kubenswrapper[4775]: I0321 05:03:32.483289 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:03:33 crc kubenswrapper[4775]: I0321 05:03:33.578241 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-85fcfb8fbb-q2k4k" event={"ID":"899f7d20-7208-419e-b0f8-36c7fbf2e841","Type":"ContainerStarted","Data":"7ee4cc8250444108e2a58c00da5799e51014a6f93fdd0496dab8bcdab841d184"} Mar 21 05:03:33 crc kubenswrapper[4775]: I0321 05:03:33.578581 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-85fcfb8fbb-q2k4k" Mar 21 05:03:33 crc kubenswrapper[4775]: I0321 05:03:33.606152 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-85fcfb8fbb-q2k4k" podStartSLOduration=2.257313755 podStartE2EDuration="5.606135892s" podCreationTimestamp="2026-03-21 05:03:28 +0000 UTC" firstStartedPulling="2026-03-21 05:03:29.181394664 +0000 UTC m=+962.157858288" lastFinishedPulling="2026-03-21 05:03:32.530216801 +0000 UTC m=+965.506680425" observedRunningTime="2026-03-21 05:03:33.602186601 +0000 UTC m=+966.578650225" watchObservedRunningTime="2026-03-21 05:03:33.606135892 +0000 UTC m=+966.582599516" Mar 21 05:03:38 crc kubenswrapper[4775]: I0321 05:03:38.740627 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-85fcfb8fbb-q2k4k" Mar 21 05:03:56 crc kubenswrapper[4775]: I0321 05:03:56.461003 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vd7j8"] Mar 21 05:03:56 crc kubenswrapper[4775]: I0321 05:03:56.462588 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vd7j8" Mar 21 05:03:56 crc kubenswrapper[4775]: I0321 05:03:56.478237 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vd7j8"] Mar 21 05:03:56 crc kubenswrapper[4775]: I0321 05:03:56.619517 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pvjj\" (UniqueName: \"kubernetes.io/projected/0f51c821-d0a6-4b50-886c-60b46cef4f4e-kube-api-access-8pvjj\") pod \"redhat-marketplace-vd7j8\" (UID: \"0f51c821-d0a6-4b50-886c-60b46cef4f4e\") " pod="openshift-marketplace/redhat-marketplace-vd7j8" Mar 21 05:03:56 crc kubenswrapper[4775]: I0321 05:03:56.619589 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f51c821-d0a6-4b50-886c-60b46cef4f4e-utilities\") pod \"redhat-marketplace-vd7j8\" (UID: \"0f51c821-d0a6-4b50-886c-60b46cef4f4e\") " pod="openshift-marketplace/redhat-marketplace-vd7j8" Mar 21 05:03:56 crc kubenswrapper[4775]: I0321 05:03:56.619611 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f51c821-d0a6-4b50-886c-60b46cef4f4e-catalog-content\") pod \"redhat-marketplace-vd7j8\" (UID: \"0f51c821-d0a6-4b50-886c-60b46cef4f4e\") " pod="openshift-marketplace/redhat-marketplace-vd7j8" Mar 21 05:03:56 crc kubenswrapper[4775]: I0321 05:03:56.721163 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f51c821-d0a6-4b50-886c-60b46cef4f4e-utilities\") pod \"redhat-marketplace-vd7j8\" (UID: \"0f51c821-d0a6-4b50-886c-60b46cef4f4e\") " pod="openshift-marketplace/redhat-marketplace-vd7j8" Mar 21 05:03:56 crc kubenswrapper[4775]: I0321 05:03:56.721232 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f51c821-d0a6-4b50-886c-60b46cef4f4e-catalog-content\") pod \"redhat-marketplace-vd7j8\" (UID: \"0f51c821-d0a6-4b50-886c-60b46cef4f4e\") " pod="openshift-marketplace/redhat-marketplace-vd7j8" Mar 21 05:03:56 crc kubenswrapper[4775]: I0321 05:03:56.721334 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pvjj\" (UniqueName: \"kubernetes.io/projected/0f51c821-d0a6-4b50-886c-60b46cef4f4e-kube-api-access-8pvjj\") pod \"redhat-marketplace-vd7j8\" (UID: \"0f51c821-d0a6-4b50-886c-60b46cef4f4e\") " pod="openshift-marketplace/redhat-marketplace-vd7j8" Mar 21 05:03:56 crc kubenswrapper[4775]: I0321 05:03:56.721744 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f51c821-d0a6-4b50-886c-60b46cef4f4e-utilities\") pod \"redhat-marketplace-vd7j8\" (UID: \"0f51c821-d0a6-4b50-886c-60b46cef4f4e\") " pod="openshift-marketplace/redhat-marketplace-vd7j8" Mar 21 05:03:56 crc kubenswrapper[4775]: I0321 05:03:56.721780 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f51c821-d0a6-4b50-886c-60b46cef4f4e-catalog-content\") pod \"redhat-marketplace-vd7j8\" (UID: \"0f51c821-d0a6-4b50-886c-60b46cef4f4e\") " pod="openshift-marketplace/redhat-marketplace-vd7j8" Mar 21 05:03:56 crc kubenswrapper[4775]: I0321 05:03:56.742632 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pvjj\" (UniqueName: \"kubernetes.io/projected/0f51c821-d0a6-4b50-886c-60b46cef4f4e-kube-api-access-8pvjj\") pod \"redhat-marketplace-vd7j8\" (UID: \"0f51c821-d0a6-4b50-886c-60b46cef4f4e\") " pod="openshift-marketplace/redhat-marketplace-vd7j8" Mar 21 05:03:56 crc kubenswrapper[4775]: I0321 05:03:56.780718 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vd7j8" Mar 21 05:03:57 crc kubenswrapper[4775]: I0321 05:03:57.302809 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vd7j8"] Mar 21 05:03:57 crc kubenswrapper[4775]: I0321 05:03:57.735088 4775 generic.go:334] "Generic (PLEG): container finished" podID="0f51c821-d0a6-4b50-886c-60b46cef4f4e" containerID="878591f3f3b4a3527080420c1c74c3cb1ee32f80a5d8b68aea2af52a45e65a6b" exitCode=0 Mar 21 05:03:57 crc kubenswrapper[4775]: I0321 05:03:57.735150 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vd7j8" event={"ID":"0f51c821-d0a6-4b50-886c-60b46cef4f4e","Type":"ContainerDied","Data":"878591f3f3b4a3527080420c1c74c3cb1ee32f80a5d8b68aea2af52a45e65a6b"} Mar 21 05:03:57 crc kubenswrapper[4775]: I0321 05:03:57.735434 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vd7j8" event={"ID":"0f51c821-d0a6-4b50-886c-60b46cef4f4e","Type":"ContainerStarted","Data":"f95851be8349176e3d371bd1b072ead3e10266f7c5f76bdcbc17b6e01db6f9b1"} Mar 21 05:03:58 crc kubenswrapper[4775]: I0321 05:03:58.743746 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vd7j8" event={"ID":"0f51c821-d0a6-4b50-886c-60b46cef4f4e","Type":"ContainerStarted","Data":"e6ba57ef1cf15fe65d271c1c5af269adfaf901ca94cd5c198057f57301b19ca3"} Mar 21 05:03:59 crc kubenswrapper[4775]: I0321 05:03:59.243513 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6nhhl"] Mar 21 05:03:59 crc kubenswrapper[4775]: I0321 05:03:59.254239 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nhhl" Mar 21 05:03:59 crc kubenswrapper[4775]: I0321 05:03:59.266662 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6nhhl"] Mar 21 05:03:59 crc kubenswrapper[4775]: I0321 05:03:59.363963 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg6ln\" (UniqueName: \"kubernetes.io/projected/694f1c66-ce56-4854-9f4c-02a42e16046d-kube-api-access-xg6ln\") pod \"community-operators-6nhhl\" (UID: \"694f1c66-ce56-4854-9f4c-02a42e16046d\") " pod="openshift-marketplace/community-operators-6nhhl" Mar 21 05:03:59 crc kubenswrapper[4775]: I0321 05:03:59.364247 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/694f1c66-ce56-4854-9f4c-02a42e16046d-utilities\") pod \"community-operators-6nhhl\" (UID: \"694f1c66-ce56-4854-9f4c-02a42e16046d\") " pod="openshift-marketplace/community-operators-6nhhl" Mar 21 05:03:59 crc kubenswrapper[4775]: I0321 05:03:59.364369 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/694f1c66-ce56-4854-9f4c-02a42e16046d-catalog-content\") pod \"community-operators-6nhhl\" (UID: \"694f1c66-ce56-4854-9f4c-02a42e16046d\") " pod="openshift-marketplace/community-operators-6nhhl" Mar 21 05:03:59 crc kubenswrapper[4775]: I0321 05:03:59.465433 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg6ln\" (UniqueName: \"kubernetes.io/projected/694f1c66-ce56-4854-9f4c-02a42e16046d-kube-api-access-xg6ln\") pod \"community-operators-6nhhl\" (UID: \"694f1c66-ce56-4854-9f4c-02a42e16046d\") " pod="openshift-marketplace/community-operators-6nhhl" Mar 21 05:03:59 crc kubenswrapper[4775]: I0321 05:03:59.465506 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/694f1c66-ce56-4854-9f4c-02a42e16046d-utilities\") pod \"community-operators-6nhhl\" (UID: \"694f1c66-ce56-4854-9f4c-02a42e16046d\") " pod="openshift-marketplace/community-operators-6nhhl" Mar 21 05:03:59 crc kubenswrapper[4775]: I0321 05:03:59.465525 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/694f1c66-ce56-4854-9f4c-02a42e16046d-catalog-content\") pod \"community-operators-6nhhl\" (UID: \"694f1c66-ce56-4854-9f4c-02a42e16046d\") " pod="openshift-marketplace/community-operators-6nhhl" Mar 21 05:03:59 crc kubenswrapper[4775]: I0321 05:03:59.465913 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/694f1c66-ce56-4854-9f4c-02a42e16046d-catalog-content\") pod \"community-operators-6nhhl\" (UID: \"694f1c66-ce56-4854-9f4c-02a42e16046d\") " pod="openshift-marketplace/community-operators-6nhhl" Mar 21 05:03:59 crc kubenswrapper[4775]: I0321 05:03:59.466171 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/694f1c66-ce56-4854-9f4c-02a42e16046d-utilities\") pod \"community-operators-6nhhl\" (UID: \"694f1c66-ce56-4854-9f4c-02a42e16046d\") " pod="openshift-marketplace/community-operators-6nhhl" Mar 21 05:03:59 crc kubenswrapper[4775]: I0321 05:03:59.484401 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg6ln\" (UniqueName: \"kubernetes.io/projected/694f1c66-ce56-4854-9f4c-02a42e16046d-kube-api-access-xg6ln\") pod \"community-operators-6nhhl\" (UID: \"694f1c66-ce56-4854-9f4c-02a42e16046d\") " pod="openshift-marketplace/community-operators-6nhhl" Mar 21 05:03:59 crc kubenswrapper[4775]: I0321 05:03:59.574877 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nhhl" Mar 21 05:03:59 crc kubenswrapper[4775]: I0321 05:03:59.757381 4775 generic.go:334] "Generic (PLEG): container finished" podID="0f51c821-d0a6-4b50-886c-60b46cef4f4e" containerID="e6ba57ef1cf15fe65d271c1c5af269adfaf901ca94cd5c198057f57301b19ca3" exitCode=0 Mar 21 05:03:59 crc kubenswrapper[4775]: I0321 05:03:59.757431 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vd7j8" event={"ID":"0f51c821-d0a6-4b50-886c-60b46cef4f4e","Type":"ContainerDied","Data":"e6ba57ef1cf15fe65d271c1c5af269adfaf901ca94cd5c198057f57301b19ca3"} Mar 21 05:03:59 crc kubenswrapper[4775]: I0321 05:03:59.968007 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6nhhl"] Mar 21 05:04:00 crc kubenswrapper[4775]: I0321 05:04:00.164458 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567824-jsc7j"] Mar 21 05:04:00 crc kubenswrapper[4775]: I0321 05:04:00.165686 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567824-jsc7j" Mar 21 05:04:00 crc kubenswrapper[4775]: I0321 05:04:00.173807 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:04:00 crc kubenswrapper[4775]: I0321 05:04:00.173897 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:04:00 crc kubenswrapper[4775]: I0321 05:04:00.173811 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:04:00 crc kubenswrapper[4775]: I0321 05:04:00.183066 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s48l\" (UniqueName: \"kubernetes.io/projected/c861c77c-15ba-4204-8603-e6093bc1a0b8-kube-api-access-7s48l\") pod \"auto-csr-approver-29567824-jsc7j\" (UID: \"c861c77c-15ba-4204-8603-e6093bc1a0b8\") " pod="openshift-infra/auto-csr-approver-29567824-jsc7j" Mar 21 05:04:00 crc kubenswrapper[4775]: I0321 05:04:00.187392 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567824-jsc7j"] Mar 21 05:04:00 crc kubenswrapper[4775]: I0321 05:04:00.284247 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s48l\" (UniqueName: \"kubernetes.io/projected/c861c77c-15ba-4204-8603-e6093bc1a0b8-kube-api-access-7s48l\") pod \"auto-csr-approver-29567824-jsc7j\" (UID: \"c861c77c-15ba-4204-8603-e6093bc1a0b8\") " pod="openshift-infra/auto-csr-approver-29567824-jsc7j" Mar 21 05:04:00 crc kubenswrapper[4775]: I0321 05:04:00.304424 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s48l\" (UniqueName: \"kubernetes.io/projected/c861c77c-15ba-4204-8603-e6093bc1a0b8-kube-api-access-7s48l\") pod \"auto-csr-approver-29567824-jsc7j\" (UID: \"c861c77c-15ba-4204-8603-e6093bc1a0b8\") " pod="openshift-infra/auto-csr-approver-29567824-jsc7j" Mar 21 05:04:00 crc kubenswrapper[4775]: I0321 05:04:00.489233 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567824-jsc7j" Mar 21 05:04:00 crc kubenswrapper[4775]: I0321 05:04:00.710789 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567824-jsc7j"] Mar 21 05:04:00 crc kubenswrapper[4775]: W0321 05:04:00.714625 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc861c77c_15ba_4204_8603_e6093bc1a0b8.slice/crio-740ad394c5df4828658e1735d663bd2c0776d52184fe75fdf22a2c8c653e68ed WatchSource:0}: Error finding container 740ad394c5df4828658e1735d663bd2c0776d52184fe75fdf22a2c8c653e68ed: Status 404 returned error can't find the container with id 740ad394c5df4828658e1735d663bd2c0776d52184fe75fdf22a2c8c653e68ed Mar 21 05:04:00 crc kubenswrapper[4775]: I0321 05:04:00.765137 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nhhl" event={"ID":"694f1c66-ce56-4854-9f4c-02a42e16046d","Type":"ContainerStarted","Data":"5780c596a9174ce034338976560a433a9f2ac7cb3aacc03fbb37bc5801d6945f"} Mar 21 05:04:00 crc kubenswrapper[4775]: I0321 05:04:00.765200 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nhhl" event={"ID":"694f1c66-ce56-4854-9f4c-02a42e16046d","Type":"ContainerStarted","Data":"3f6120ae96f0b10b5bf76bd3efa458fabb2c0680a195b83d1f67570f61231bb6"} Mar 21 05:04:00 crc kubenswrapper[4775]: I0321 05:04:00.767008 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567824-jsc7j" event={"ID":"c861c77c-15ba-4204-8603-e6093bc1a0b8","Type":"ContainerStarted","Data":"740ad394c5df4828658e1735d663bd2c0776d52184fe75fdf22a2c8c653e68ed"} Mar 21 05:04:01 crc kubenswrapper[4775]: I0321 05:04:01.775586 4775 generic.go:334] "Generic (PLEG): container finished" podID="694f1c66-ce56-4854-9f4c-02a42e16046d" containerID="5780c596a9174ce034338976560a433a9f2ac7cb3aacc03fbb37bc5801d6945f" exitCode=0 Mar 21 05:04:01 crc kubenswrapper[4775]: I0321 05:04:01.775643 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nhhl" event={"ID":"694f1c66-ce56-4854-9f4c-02a42e16046d","Type":"ContainerDied","Data":"5780c596a9174ce034338976560a433a9f2ac7cb3aacc03fbb37bc5801d6945f"} Mar 21 05:04:01 crc kubenswrapper[4775]: I0321 05:04:01.778336 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vd7j8" event={"ID":"0f51c821-d0a6-4b50-886c-60b46cef4f4e","Type":"ContainerStarted","Data":"81bc0fdd8eb47b55fa064c84d2bb340801a9116fe70f3e09b004ec5b7bda0d6c"} Mar 21 05:04:01 crc kubenswrapper[4775]: I0321 05:04:01.813168 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vd7j8" podStartSLOduration=2.781509761 podStartE2EDuration="5.813100508s" podCreationTimestamp="2026-03-21 05:03:56 +0000 UTC" firstStartedPulling="2026-03-21 05:03:57.737308118 +0000 UTC m=+990.713771742" lastFinishedPulling="2026-03-21 05:04:00.768898865 +0000 UTC m=+993.745362489" observedRunningTime="2026-03-21 05:04:01.810052581 +0000 UTC m=+994.786516195" watchObservedRunningTime="2026-03-21 05:04:01.813100508 +0000 UTC m=+994.789564132" Mar 21 05:04:02 crc kubenswrapper[4775]: I0321 05:04:02.482777 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:04:02 crc kubenswrapper[4775]: I0321 05:04:02.482832 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:04:02 crc kubenswrapper[4775]: I0321 05:04:02.785742 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nhhl" event={"ID":"694f1c66-ce56-4854-9f4c-02a42e16046d","Type":"ContainerStarted","Data":"428647d2139b15f3ab85f6734bc25c0214e38d663b4a1c7cb9b69a834fdbe970"} Mar 21 05:04:02 crc kubenswrapper[4775]: I0321 05:04:02.787650 4775 generic.go:334] "Generic (PLEG): container finished" podID="c861c77c-15ba-4204-8603-e6093bc1a0b8" containerID="ca9c08541581bc07e396b682f2ba269a753d458f0a5cc3ff5f26ace1828dea81" exitCode=0 Mar 21 05:04:02 crc kubenswrapper[4775]: I0321 05:04:02.787696 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567824-jsc7j" event={"ID":"c861c77c-15ba-4204-8603-e6093bc1a0b8","Type":"ContainerDied","Data":"ca9c08541581bc07e396b682f2ba269a753d458f0a5cc3ff5f26ace1828dea81"} Mar 21 05:04:03 crc kubenswrapper[4775]: I0321 05:04:03.794449 4775 generic.go:334] "Generic (PLEG): container finished" podID="694f1c66-ce56-4854-9f4c-02a42e16046d" containerID="428647d2139b15f3ab85f6734bc25c0214e38d663b4a1c7cb9b69a834fdbe970" exitCode=0 Mar 21 05:04:03 crc kubenswrapper[4775]: I0321 05:04:03.795256 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nhhl" event={"ID":"694f1c66-ce56-4854-9f4c-02a42e16046d","Type":"ContainerDied","Data":"428647d2139b15f3ab85f6734bc25c0214e38d663b4a1c7cb9b69a834fdbe970"} Mar 21 05:04:04 crc kubenswrapper[4775]: I0321 05:04:04.050861 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567824-jsc7j" Mar 21 05:04:04 crc kubenswrapper[4775]: I0321 05:04:04.245674 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7s48l\" (UniqueName: \"kubernetes.io/projected/c861c77c-15ba-4204-8603-e6093bc1a0b8-kube-api-access-7s48l\") pod \"c861c77c-15ba-4204-8603-e6093bc1a0b8\" (UID: \"c861c77c-15ba-4204-8603-e6093bc1a0b8\") " Mar 21 05:04:04 crc kubenswrapper[4775]: I0321 05:04:04.252671 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c861c77c-15ba-4204-8603-e6093bc1a0b8-kube-api-access-7s48l" (OuterVolumeSpecName: "kube-api-access-7s48l") pod "c861c77c-15ba-4204-8603-e6093bc1a0b8" (UID: "c861c77c-15ba-4204-8603-e6093bc1a0b8"). InnerVolumeSpecName "kube-api-access-7s48l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:04:04 crc kubenswrapper[4775]: I0321 05:04:04.347362 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7s48l\" (UniqueName: \"kubernetes.io/projected/c861c77c-15ba-4204-8603-e6093bc1a0b8-kube-api-access-7s48l\") on node \"crc\" DevicePath \"\"" Mar 21 05:04:04 crc kubenswrapper[4775]: I0321 05:04:04.803160 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nhhl" event={"ID":"694f1c66-ce56-4854-9f4c-02a42e16046d","Type":"ContainerStarted","Data":"bfa25ad839c3791aa796fdbb1ca916c818ffc84d5a2bd52f3d19c4a9886bd0e1"} Mar 21 05:04:04 crc kubenswrapper[4775]: I0321 05:04:04.805566 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567824-jsc7j" event={"ID":"c861c77c-15ba-4204-8603-e6093bc1a0b8","Type":"ContainerDied","Data":"740ad394c5df4828658e1735d663bd2c0776d52184fe75fdf22a2c8c653e68ed"} Mar 21 05:04:04 crc kubenswrapper[4775]: I0321 05:04:04.805604 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="740ad394c5df4828658e1735d663bd2c0776d52184fe75fdf22a2c8c653e68ed" Mar 21 05:04:04 crc kubenswrapper[4775]: I0321 05:04:04.805620 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567824-jsc7j" Mar 21 05:04:04 crc kubenswrapper[4775]: I0321 05:04:04.835873 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6nhhl" podStartSLOduration=3.344814295 podStartE2EDuration="5.835853343s" podCreationTimestamp="2026-03-21 05:03:59 +0000 UTC" firstStartedPulling="2026-03-21 05:04:01.776923962 +0000 UTC m=+994.753387596" lastFinishedPulling="2026-03-21 05:04:04.26796302 +0000 UTC m=+997.244426644" observedRunningTime="2026-03-21 05:04:04.835571075 +0000 UTC m=+997.812034699" watchObservedRunningTime="2026-03-21 05:04:04.835853343 +0000 UTC m=+997.812316967" Mar 21 05:04:05 crc kubenswrapper[4775]: I0321 05:04:05.139973 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567818-42hv5"] Mar 21 05:04:05 crc kubenswrapper[4775]: I0321 05:04:05.145719 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567818-42hv5"] Mar 21 05:04:05 crc kubenswrapper[4775]: I0321 05:04:05.669326 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40cf5461-a1ca-40cb-9ed3-9c2de90faad3" path="/var/lib/kubelet/pods/40cf5461-a1ca-40cb-9ed3-9c2de90faad3/volumes" Mar 21 05:04:06 crc kubenswrapper[4775]: I0321 05:04:06.781428 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vd7j8" Mar 21 05:04:06 crc kubenswrapper[4775]: I0321 05:04:06.782330 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vd7j8" Mar 21 05:04:06 crc kubenswrapper[4775]: I0321 05:04:06.843411 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vd7j8" Mar 21 05:04:07 crc kubenswrapper[4775]: I0321 05:04:07.869056 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vd7j8" Mar 21 05:04:09 crc kubenswrapper[4775]: I0321 05:04:09.034987 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vd7j8"] Mar 21 05:04:09 crc kubenswrapper[4775]: I0321 05:04:09.575191 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6nhhl" Mar 21 05:04:09 crc kubenswrapper[4775]: I0321 05:04:09.575300 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6nhhl" Mar 21 05:04:09 crc kubenswrapper[4775]: I0321 05:04:09.620247 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6nhhl" Mar 21 05:04:09 crc kubenswrapper[4775]: I0321 05:04:09.834031 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vd7j8" podUID="0f51c821-d0a6-4b50-886c-60b46cef4f4e" containerName="registry-server" containerID="cri-o://81bc0fdd8eb47b55fa064c84d2bb340801a9116fe70f3e09b004ec5b7bda0d6c" gracePeriod=2 Mar 21 05:04:09 crc kubenswrapper[4775]: I0321 05:04:09.877873 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6nhhl" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.271332 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vd7j8" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.328173 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f51c821-d0a6-4b50-886c-60b46cef4f4e-catalog-content\") pod \"0f51c821-d0a6-4b50-886c-60b46cef4f4e\" (UID: \"0f51c821-d0a6-4b50-886c-60b46cef4f4e\") " Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.328221 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pvjj\" (UniqueName: \"kubernetes.io/projected/0f51c821-d0a6-4b50-886c-60b46cef4f4e-kube-api-access-8pvjj\") pod \"0f51c821-d0a6-4b50-886c-60b46cef4f4e\" (UID: \"0f51c821-d0a6-4b50-886c-60b46cef4f4e\") " Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.328267 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f51c821-d0a6-4b50-886c-60b46cef4f4e-utilities\") pod \"0f51c821-d0a6-4b50-886c-60b46cef4f4e\" (UID: \"0f51c821-d0a6-4b50-886c-60b46cef4f4e\") " Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.329198 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f51c821-d0a6-4b50-886c-60b46cef4f4e-utilities" (OuterVolumeSpecName: "utilities") pod "0f51c821-d0a6-4b50-886c-60b46cef4f4e" (UID: "0f51c821-d0a6-4b50-886c-60b46cef4f4e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.347616 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f51c821-d0a6-4b50-886c-60b46cef4f4e-kube-api-access-8pvjj" (OuterVolumeSpecName: "kube-api-access-8pvjj") pod "0f51c821-d0a6-4b50-886c-60b46cef4f4e" (UID: "0f51c821-d0a6-4b50-886c-60b46cef4f4e"). InnerVolumeSpecName "kube-api-access-8pvjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.360373 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f51c821-d0a6-4b50-886c-60b46cef4f4e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f51c821-d0a6-4b50-886c-60b46cef4f4e" (UID: "0f51c821-d0a6-4b50-886c-60b46cef4f4e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.430054 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f51c821-d0a6-4b50-886c-60b46cef4f4e-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.430093 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f51c821-d0a6-4b50-886c-60b46cef4f4e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.430107 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pvjj\" (UniqueName: \"kubernetes.io/projected/0f51c821-d0a6-4b50-886c-60b46cef4f4e-kube-api-access-8pvjj\") on node \"crc\" DevicePath \"\"" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.842976 4775 generic.go:334] "Generic (PLEG): container finished" podID="0f51c821-d0a6-4b50-886c-60b46cef4f4e" containerID="81bc0fdd8eb47b55fa064c84d2bb340801a9116fe70f3e09b004ec5b7bda0d6c" exitCode=0 Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.843046 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vd7j8" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.843087 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vd7j8" event={"ID":"0f51c821-d0a6-4b50-886c-60b46cef4f4e","Type":"ContainerDied","Data":"81bc0fdd8eb47b55fa064c84d2bb340801a9116fe70f3e09b004ec5b7bda0d6c"} Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.843140 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vd7j8" event={"ID":"0f51c821-d0a6-4b50-886c-60b46cef4f4e","Type":"ContainerDied","Data":"f95851be8349176e3d371bd1b072ead3e10266f7c5f76bdcbc17b6e01db6f9b1"} Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.843162 4775 scope.go:117] "RemoveContainer" containerID="81bc0fdd8eb47b55fa064c84d2bb340801a9116fe70f3e09b004ec5b7bda0d6c" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.869286 4775 scope.go:117] "RemoveContainer" containerID="e6ba57ef1cf15fe65d271c1c5af269adfaf901ca94cd5c198057f57301b19ca3" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.894252 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vd7j8"] Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.900648 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vd7j8"] Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.904087 4775 scope.go:117] "RemoveContainer" containerID="878591f3f3b4a3527080420c1c74c3cb1ee32f80a5d8b68aea2af52a45e65a6b" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.922812 4775 scope.go:117] "RemoveContainer" containerID="81bc0fdd8eb47b55fa064c84d2bb340801a9116fe70f3e09b004ec5b7bda0d6c" Mar 21 05:04:10 crc kubenswrapper[4775]: E0321 05:04:10.923232 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81bc0fdd8eb47b55fa064c84d2bb340801a9116fe70f3e09b004ec5b7bda0d6c\": container with ID starting with 81bc0fdd8eb47b55fa064c84d2bb340801a9116fe70f3e09b004ec5b7bda0d6c not found: ID does not exist" containerID="81bc0fdd8eb47b55fa064c84d2bb340801a9116fe70f3e09b004ec5b7bda0d6c" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.923261 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81bc0fdd8eb47b55fa064c84d2bb340801a9116fe70f3e09b004ec5b7bda0d6c"} err="failed to get container status \"81bc0fdd8eb47b55fa064c84d2bb340801a9116fe70f3e09b004ec5b7bda0d6c\": rpc error: code = NotFound desc = could not find container \"81bc0fdd8eb47b55fa064c84d2bb340801a9116fe70f3e09b004ec5b7bda0d6c\": container with ID starting with 81bc0fdd8eb47b55fa064c84d2bb340801a9116fe70f3e09b004ec5b7bda0d6c not found: ID does not exist" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.923281 4775 scope.go:117] "RemoveContainer" containerID="e6ba57ef1cf15fe65d271c1c5af269adfaf901ca94cd5c198057f57301b19ca3" Mar 21 05:04:10 crc kubenswrapper[4775]: E0321 05:04:10.923475 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6ba57ef1cf15fe65d271c1c5af269adfaf901ca94cd5c198057f57301b19ca3\": container with ID starting with e6ba57ef1cf15fe65d271c1c5af269adfaf901ca94cd5c198057f57301b19ca3 not found: ID does not exist" containerID="e6ba57ef1cf15fe65d271c1c5af269adfaf901ca94cd5c198057f57301b19ca3" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.923496 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6ba57ef1cf15fe65d271c1c5af269adfaf901ca94cd5c198057f57301b19ca3"} err="failed to get container status \"e6ba57ef1cf15fe65d271c1c5af269adfaf901ca94cd5c198057f57301b19ca3\": rpc error: code = NotFound desc = could not find container \"e6ba57ef1cf15fe65d271c1c5af269adfaf901ca94cd5c198057f57301b19ca3\": container with ID starting with e6ba57ef1cf15fe65d271c1c5af269adfaf901ca94cd5c198057f57301b19ca3 not found: ID does not exist" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.923513 4775 scope.go:117] "RemoveContainer" containerID="878591f3f3b4a3527080420c1c74c3cb1ee32f80a5d8b68aea2af52a45e65a6b" Mar 21 05:04:10 crc kubenswrapper[4775]: E0321 05:04:10.923686 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"878591f3f3b4a3527080420c1c74c3cb1ee32f80a5d8b68aea2af52a45e65a6b\": container with ID starting with 878591f3f3b4a3527080420c1c74c3cb1ee32f80a5d8b68aea2af52a45e65a6b not found: ID does not exist" containerID="878591f3f3b4a3527080420c1c74c3cb1ee32f80a5d8b68aea2af52a45e65a6b" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.923706 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"878591f3f3b4a3527080420c1c74c3cb1ee32f80a5d8b68aea2af52a45e65a6b"} err="failed to get container status \"878591f3f3b4a3527080420c1c74c3cb1ee32f80a5d8b68aea2af52a45e65a6b\": rpc error: code = NotFound desc = could not find container \"878591f3f3b4a3527080420c1c74c3cb1ee32f80a5d8b68aea2af52a45e65a6b\": container with ID starting with 878591f3f3b4a3527080420c1c74c3cb1ee32f80a5d8b68aea2af52a45e65a6b not found: ID does not exist" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.998185 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-twhxx"] Mar 21 05:04:10 crc kubenswrapper[4775]: E0321 05:04:10.998422 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f51c821-d0a6-4b50-886c-60b46cef4f4e" containerName="extract-content" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.998437 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f51c821-d0a6-4b50-886c-60b46cef4f4e" containerName="extract-content" Mar 21 05:04:10 crc kubenswrapper[4775]: E0321 05:04:10.998448 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c861c77c-15ba-4204-8603-e6093bc1a0b8" containerName="oc" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.998454 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c861c77c-15ba-4204-8603-e6093bc1a0b8" containerName="oc" Mar 21 05:04:10 crc kubenswrapper[4775]: E0321 05:04:10.998464 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f51c821-d0a6-4b50-886c-60b46cef4f4e" containerName="registry-server" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.998469 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f51c821-d0a6-4b50-886c-60b46cef4f4e" containerName="registry-server" Mar 21 05:04:10 crc kubenswrapper[4775]: E0321 05:04:10.998489 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f51c821-d0a6-4b50-886c-60b46cef4f4e" containerName="extract-utilities" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.998495 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f51c821-d0a6-4b50-886c-60b46cef4f4e" containerName="extract-utilities" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.998637 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c861c77c-15ba-4204-8603-e6093bc1a0b8" containerName="oc" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.998651 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f51c821-d0a6-4b50-886c-60b46cef4f4e" containerName="registry-server" Mar 21 05:04:10 crc kubenswrapper[4775]: I0321 05:04:10.999042 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-twhxx" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.002042 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-n6256" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.007754 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-dn22m"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.009060 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-dn22m" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.011563 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-bjdr6" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.018895 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-twhxx"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.028128 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-dn22m"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.036834 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l564\" (UniqueName: \"kubernetes.io/projected/94e1507b-be6c-4ecf-99e5-2bdcd2cc0cef-kube-api-access-2l564\") pod \"barbican-operator-controller-manager-59bc569d95-twhxx\" (UID: \"94e1507b-be6c-4ecf-99e5-2bdcd2cc0cef\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-twhxx" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.037127 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbgp7\" (UniqueName: \"kubernetes.io/projected/9fe71acc-7d35-4d4b-ac69-e193d3f39028-kube-api-access-fbgp7\") pod \"cinder-operator-controller-manager-8d58dc466-dn22m\" (UID: \"9fe71acc-7d35-4d4b-ac69-e193d3f39028\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-dn22m" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.042204 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-lxvtw"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.042986 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-lxvtw" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.047975 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-8v77x" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.067167 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-lzvsq"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.067893 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-lzvsq" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.072419 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-qrwxc" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.082031 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-lxvtw"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.094947 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-7bdbc"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.095936 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-7bdbc" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.103345 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-rnw8r" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.104345 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-m4gqz"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.105213 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-m4gqz" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.107090 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-vdlb4" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.116928 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-lzvsq"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.119843 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-m4gqz"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.143693 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l564\" (UniqueName: \"kubernetes.io/projected/94e1507b-be6c-4ecf-99e5-2bdcd2cc0cef-kube-api-access-2l564\") pod \"barbican-operator-controller-manager-59bc569d95-twhxx\" (UID: \"94e1507b-be6c-4ecf-99e5-2bdcd2cc0cef\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-twhxx" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.143740 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbgp7\" (UniqueName: \"kubernetes.io/projected/9fe71acc-7d35-4d4b-ac69-e193d3f39028-kube-api-access-fbgp7\") pod \"cinder-operator-controller-manager-8d58dc466-dn22m\" (UID: \"9fe71acc-7d35-4d4b-ac69-e193d3f39028\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-dn22m" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.143773 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb4js\" (UniqueName: \"kubernetes.io/projected/87dcea67-7f65-46a6-996b-3985bf1b5171-kube-api-access-sb4js\") pod \"designate-operator-controller-manager-588d4d986b-lxvtw\" (UID: \"87dcea67-7f65-46a6-996b-3985bf1b5171\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-lxvtw" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.143805 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftw2s\" (UniqueName: \"kubernetes.io/projected/01ed348d-8a8a-4717-ba0d-1944b3f1c081-kube-api-access-ftw2s\") pod \"horizon-operator-controller-manager-8464cc45fb-7bdbc\" (UID: \"01ed348d-8a8a-4717-ba0d-1944b3f1c081\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-7bdbc" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.143826 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h94bs\" (UniqueName: \"kubernetes.io/projected/0e83601c-758c-4f12-b745-bb68b0c4904f-kube-api-access-h94bs\") pod \"heat-operator-controller-manager-67dd5f86f5-m4gqz\" (UID: \"0e83601c-758c-4f12-b745-bb68b0c4904f\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-m4gqz" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.143874 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77nmg\" (UniqueName: \"kubernetes.io/projected/80932361-6406-48dd-9e4b-4e9c27813f68-kube-api-access-77nmg\") pod \"glance-operator-controller-manager-79df6bcc97-lzvsq\" (UID: \"80932361-6406-48dd-9e4b-4e9c27813f68\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-lzvsq" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.144349 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.145056 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.148010 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.148231 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-fdr25" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.170370 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-9g8pv"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.171512 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-9g8pv" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.181879 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-7bdbc"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.182057 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-7pzrk" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.183431 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l564\" (UniqueName: \"kubernetes.io/projected/94e1507b-be6c-4ecf-99e5-2bdcd2cc0cef-kube-api-access-2l564\") pod \"barbican-operator-controller-manager-59bc569d95-twhxx\" (UID: \"94e1507b-be6c-4ecf-99e5-2bdcd2cc0cef\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-twhxx" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.191682 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbgp7\" (UniqueName: \"kubernetes.io/projected/9fe71acc-7d35-4d4b-ac69-e193d3f39028-kube-api-access-fbgp7\") pod \"cinder-operator-controller-manager-8d58dc466-dn22m\" (UID: \"9fe71acc-7d35-4d4b-ac69-e193d3f39028\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-dn22m" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.194166 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.200435 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-9g8pv"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.205231 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-c7pjw"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.205941 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c7pjw" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.211742 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-2vrnb" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.212048 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-jq88h"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.212763 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-55f864c847-jq88h" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.221654 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-jq88h"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.223187 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-b6stv" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.242150 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-c7pjw"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.244837 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftw2s\" (UniqueName: \"kubernetes.io/projected/01ed348d-8a8a-4717-ba0d-1944b3f1c081-kube-api-access-ftw2s\") pod \"horizon-operator-controller-manager-8464cc45fb-7bdbc\" (UID: \"01ed348d-8a8a-4717-ba0d-1944b3f1c081\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-7bdbc" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.245010 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h94bs\" (UniqueName: \"kubernetes.io/projected/0e83601c-758c-4f12-b745-bb68b0c4904f-kube-api-access-h94bs\") pod \"heat-operator-controller-manager-67dd5f86f5-m4gqz\" (UID: \"0e83601c-758c-4f12-b745-bb68b0c4904f\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-m4gqz" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.247044 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77nmg\" (UniqueName: \"kubernetes.io/projected/80932361-6406-48dd-9e4b-4e9c27813f68-kube-api-access-77nmg\") pod \"glance-operator-controller-manager-79df6bcc97-lzvsq\" (UID: \"80932361-6406-48dd-9e4b-4e9c27813f68\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-lzvsq" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.247164 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbqp8\" (UniqueName: \"kubernetes.io/projected/0a66456f-7860-4dc1-9c1c-0db69ddcc800-kube-api-access-gbqp8\") pod \"infra-operator-controller-manager-65f65cc49c-2mgp8\" (UID: \"0a66456f-7860-4dc1-9c1c-0db69ddcc800\") " pod="openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.247263 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8hqs\" (UniqueName: \"kubernetes.io/projected/20c78c73-daf5-481e-a4ac-62de73b5969e-kube-api-access-f8hqs\") pod \"manila-operator-controller-manager-55f864c847-jq88h\" (UID: \"20c78c73-daf5-481e-a4ac-62de73b5969e\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-jq88h" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.247381 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66ql6\" (UniqueName: \"kubernetes.io/projected/49a3c6c7-9e86-495d-8d1a-486d6bfbbbdd-kube-api-access-66ql6\") pod \"keystone-operator-controller-manager-768b96df4c-c7pjw\" (UID: \"49a3c6c7-9e86-495d-8d1a-486d6bfbbbdd\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c7pjw" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.247493 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb4js\" (UniqueName: \"kubernetes.io/projected/87dcea67-7f65-46a6-996b-3985bf1b5171-kube-api-access-sb4js\") pod \"designate-operator-controller-manager-588d4d986b-lxvtw\" (UID: \"87dcea67-7f65-46a6-996b-3985bf1b5171\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-lxvtw" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.247557 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert\") pod \"infra-operator-controller-manager-65f65cc49c-2mgp8\" (UID: \"0a66456f-7860-4dc1-9c1c-0db69ddcc800\") " pod="openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.247578 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpz8n\" (UniqueName: \"kubernetes.io/projected/f00c8c4b-874f-45ec-8a1a-e0834b3fc252-kube-api-access-gpz8n\") pod \"ironic-operator-controller-manager-6f787dddc9-9g8pv\" (UID: \"f00c8c4b-874f-45ec-8a1a-e0834b3fc252\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-9g8pv" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.249298 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hf688"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.250312 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hf688" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.252543 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-f97t2" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.256769 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-4tbvg"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.257729 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-767865f676-4tbvg" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.264163 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hf688"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.270506 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-rlj7t" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.279941 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-4tbvg"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.281325 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftw2s\" (UniqueName: \"kubernetes.io/projected/01ed348d-8a8a-4717-ba0d-1944b3f1c081-kube-api-access-ftw2s\") pod \"horizon-operator-controller-manager-8464cc45fb-7bdbc\" (UID: \"01ed348d-8a8a-4717-ba0d-1944b3f1c081\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-7bdbc" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.288486 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77nmg\" (UniqueName: \"kubernetes.io/projected/80932361-6406-48dd-9e4b-4e9c27813f68-kube-api-access-77nmg\") pod \"glance-operator-controller-manager-79df6bcc97-lzvsq\" (UID: \"80932361-6406-48dd-9e4b-4e9c27813f68\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-lzvsq" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.304886 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h94bs\" (UniqueName: \"kubernetes.io/projected/0e83601c-758c-4f12-b745-bb68b0c4904f-kube-api-access-h94bs\") pod \"heat-operator-controller-manager-67dd5f86f5-m4gqz\" (UID: \"0e83601c-758c-4f12-b745-bb68b0c4904f\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-m4gqz" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.337857 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb4js\" (UniqueName: \"kubernetes.io/projected/87dcea67-7f65-46a6-996b-3985bf1b5171-kube-api-access-sb4js\") pod \"designate-operator-controller-manager-588d4d986b-lxvtw\" (UID: \"87dcea67-7f65-46a6-996b-3985bf1b5171\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-lxvtw" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.338133 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-twhxx" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.338303 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-dn22m" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.339302 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-tss7r"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.362881 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tss7r" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.375840 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-ccm9j" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.376493 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-lxvtw" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.376721 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbqp8\" (UniqueName: \"kubernetes.io/projected/0a66456f-7860-4dc1-9c1c-0db69ddcc800-kube-api-access-gbqp8\") pod \"infra-operator-controller-manager-65f65cc49c-2mgp8\" (UID: \"0a66456f-7860-4dc1-9c1c-0db69ddcc800\") " pod="openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.389413 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8hqs\" (UniqueName: \"kubernetes.io/projected/20c78c73-daf5-481e-a4ac-62de73b5969e-kube-api-access-f8hqs\") pod \"manila-operator-controller-manager-55f864c847-jq88h\" (UID: \"20c78c73-daf5-481e-a4ac-62de73b5969e\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-jq88h" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.389513 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66ql6\" (UniqueName: \"kubernetes.io/projected/49a3c6c7-9e86-495d-8d1a-486d6bfbbbdd-kube-api-access-66ql6\") pod \"keystone-operator-controller-manager-768b96df4c-c7pjw\" (UID: \"49a3c6c7-9e86-495d-8d1a-486d6bfbbbdd\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c7pjw" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.389561 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert\") pod \"infra-operator-controller-manager-65f65cc49c-2mgp8\" (UID: \"0a66456f-7860-4dc1-9c1c-0db69ddcc800\") " pod="openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.389589 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpz8n\" (UniqueName: \"kubernetes.io/projected/f00c8c4b-874f-45ec-8a1a-e0834b3fc252-kube-api-access-gpz8n\") pod \"ironic-operator-controller-manager-6f787dddc9-9g8pv\" (UID: \"f00c8c4b-874f-45ec-8a1a-e0834b3fc252\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-9g8pv" Mar 21 05:04:11 crc kubenswrapper[4775]: E0321 05:04:11.390722 4775 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 21 05:04:11 crc kubenswrapper[4775]: E0321 05:04:11.390828 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert podName:0a66456f-7860-4dc1-9c1c-0db69ddcc800 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:11.890776544 +0000 UTC m=+1004.867240168 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert") pod "infra-operator-controller-manager-65f65cc49c-2mgp8" (UID: "0a66456f-7860-4dc1-9c1c-0db69ddcc800") : secret "infra-operator-webhook-server-cert" not found Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.392581 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-lzvsq" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.423465 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66ql6\" (UniqueName: \"kubernetes.io/projected/49a3c6c7-9e86-495d-8d1a-486d6bfbbbdd-kube-api-access-66ql6\") pod \"keystone-operator-controller-manager-768b96df4c-c7pjw\" (UID: \"49a3c6c7-9e86-495d-8d1a-486d6bfbbbdd\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c7pjw" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.423686 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbqp8\" (UniqueName: \"kubernetes.io/projected/0a66456f-7860-4dc1-9c1c-0db69ddcc800-kube-api-access-gbqp8\") pod \"infra-operator-controller-manager-65f65cc49c-2mgp8\" (UID: \"0a66456f-7860-4dc1-9c1c-0db69ddcc800\") " pod="openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.423913 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-7bdbc" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.429913 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-67scw"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.436437 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-67scw" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.436638 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpz8n\" (UniqueName: \"kubernetes.io/projected/f00c8c4b-874f-45ec-8a1a-e0834b3fc252-kube-api-access-gpz8n\") pod \"ironic-operator-controller-manager-6f787dddc9-9g8pv\" (UID: \"f00c8c4b-874f-45ec-8a1a-e0834b3fc252\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-9g8pv" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.436827 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8hqs\" (UniqueName: \"kubernetes.io/projected/20c78c73-daf5-481e-a4ac-62de73b5969e-kube-api-access-f8hqs\") pod \"manila-operator-controller-manager-55f864c847-jq88h\" (UID: \"20c78c73-daf5-481e-a4ac-62de73b5969e\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-jq88h" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.437146 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-m4gqz" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.440923 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-hvcgr" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.471313 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-tss7r"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.494636 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-67scw"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.504095 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zv76\" (UniqueName: \"kubernetes.io/projected/9bfe7d25-53ea-484d-a481-0ea04ee2b8a8-kube-api-access-8zv76\") pod \"neutron-operator-controller-manager-767865f676-4tbvg\" (UID: \"9bfe7d25-53ea-484d-a481-0ea04ee2b8a8\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-4tbvg" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.504221 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgdjp\" (UniqueName: \"kubernetes.io/projected/02b6af47-2c06-480b-a838-2d742efa1045-kube-api-access-bgdjp\") pod \"nova-operator-controller-manager-5d488d59fb-tss7r\" (UID: \"02b6af47-2c06-480b-a838-2d742efa1045\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tss7r" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.504271 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84jdm\" (UniqueName: \"kubernetes.io/projected/ec38d53e-6fe4-41b5-8548-e49fadd9d6bf-kube-api-access-84jdm\") pod \"mariadb-operator-controller-manager-67ccfc9778-hf688\" (UID: \"ec38d53e-6fe4-41b5-8548-e49fadd9d6bf\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hf688" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.504315 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tstv\" (UniqueName: \"kubernetes.io/projected/745c79b1-1bcf-4c0f-82ee-a26cbba46d48-kube-api-access-4tstv\") pod \"octavia-operator-controller-manager-5b9f45d989-67scw\" (UID: \"745c79b1-1bcf-4c0f-82ee-a26cbba46d48\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-67scw" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.525676 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.526462 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.528838 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-cthp5" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.529066 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.533465 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-jj4pt"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.534685 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5784578c99-jj4pt" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.537112 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-qq45w" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.544382 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-9g8pv" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.569379 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-jj4pt"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.576434 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c7pjw" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.576925 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-ctm9h"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.578804 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-884679f54-ctm9h" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.582836 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-sdwdh" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.588421 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.595239 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-ctm9h"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.605100 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zv76\" (UniqueName: \"kubernetes.io/projected/9bfe7d25-53ea-484d-a481-0ea04ee2b8a8-kube-api-access-8zv76\") pod \"neutron-operator-controller-manager-767865f676-4tbvg\" (UID: \"9bfe7d25-53ea-484d-a481-0ea04ee2b8a8\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-4tbvg" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.605179 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxgwb\" (UniqueName: \"kubernetes.io/projected/898b32c5-9f21-4fba-90c5-a333f36addf2-kube-api-access-cxgwb\") pod \"placement-operator-controller-manager-5784578c99-jj4pt\" (UID: \"898b32c5-9f21-4fba-90c5-a333f36addf2\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-jj4pt" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.605215 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgdjp\" (UniqueName: \"kubernetes.io/projected/02b6af47-2c06-480b-a838-2d742efa1045-kube-api-access-bgdjp\") pod \"nova-operator-controller-manager-5d488d59fb-tss7r\" (UID: \"02b6af47-2c06-480b-a838-2d742efa1045\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tss7r" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.605241 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrnmj\" (UniqueName: \"kubernetes.io/projected/3035739a-202f-4794-bb4f-ae2342a96441-kube-api-access-jrnmj\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rwq59\" (UID: \"3035739a-202f-4794-bb4f-ae2342a96441\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.605271 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rwq59\" (UID: \"3035739a-202f-4794-bb4f-ae2342a96441\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.605291 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84jdm\" (UniqueName: \"kubernetes.io/projected/ec38d53e-6fe4-41b5-8548-e49fadd9d6bf-kube-api-access-84jdm\") pod \"mariadb-operator-controller-manager-67ccfc9778-hf688\" (UID: \"ec38d53e-6fe4-41b5-8548-e49fadd9d6bf\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hf688" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.605317 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tstv\" (UniqueName: \"kubernetes.io/projected/745c79b1-1bcf-4c0f-82ee-a26cbba46d48-kube-api-access-4tstv\") pod \"octavia-operator-controller-manager-5b9f45d989-67scw\" (UID: \"745c79b1-1bcf-4c0f-82ee-a26cbba46d48\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-67scw" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.605337 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7zpk\" (UniqueName: \"kubernetes.io/projected/6eeb04ad-7251-488c-bd52-b2f14f6fb68b-kube-api-access-x7zpk\") pod \"ovn-operator-controller-manager-884679f54-ctm9h\" (UID: \"6eeb04ad-7251-488c-bd52-b2f14f6fb68b\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-ctm9h" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.627569 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgdjp\" (UniqueName: \"kubernetes.io/projected/02b6af47-2c06-480b-a838-2d742efa1045-kube-api-access-bgdjp\") pod \"nova-operator-controller-manager-5d488d59fb-tss7r\" (UID: \"02b6af47-2c06-480b-a838-2d742efa1045\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tss7r" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.627744 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-lqmgv"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.629066 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-c674c5965-lqmgv" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.631865 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-8mtqb" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.631903 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84jdm\" (UniqueName: \"kubernetes.io/projected/ec38d53e-6fe4-41b5-8548-e49fadd9d6bf-kube-api-access-84jdm\") pod \"mariadb-operator-controller-manager-67ccfc9778-hf688\" (UID: \"ec38d53e-6fe4-41b5-8548-e49fadd9d6bf\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hf688" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.633659 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zv76\" (UniqueName: \"kubernetes.io/projected/9bfe7d25-53ea-484d-a481-0ea04ee2b8a8-kube-api-access-8zv76\") pod \"neutron-operator-controller-manager-767865f676-4tbvg\" (UID: \"9bfe7d25-53ea-484d-a481-0ea04ee2b8a8\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-4tbvg" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.633721 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tstv\" (UniqueName: \"kubernetes.io/projected/745c79b1-1bcf-4c0f-82ee-a26cbba46d48-kube-api-access-4tstv\") pod \"octavia-operator-controller-manager-5b9f45d989-67scw\" (UID: \"745c79b1-1bcf-4c0f-82ee-a26cbba46d48\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-67scw" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.647971 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-lqmgv"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.702670 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f51c821-d0a6-4b50-886c-60b46cef4f4e" path="/var/lib/kubelet/pods/0f51c821-d0a6-4b50-886c-60b46cef4f4e/volumes" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.704504 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-9s82h"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.708170 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-9s82h"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.708923 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-9s82h" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.714210 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-55f864c847-jq88h" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.715599 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-czmtg" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.716078 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7zpk\" (UniqueName: \"kubernetes.io/projected/6eeb04ad-7251-488c-bd52-b2f14f6fb68b-kube-api-access-x7zpk\") pod \"ovn-operator-controller-manager-884679f54-ctm9h\" (UID: \"6eeb04ad-7251-488c-bd52-b2f14f6fb68b\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-ctm9h" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.716150 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g8sv\" (UniqueName: \"kubernetes.io/projected/5968f1d9-f4e0-4c67-923e-2494e15c4088-kube-api-access-6g8sv\") pod \"swift-operator-controller-manager-c674c5965-lqmgv\" (UID: \"5968f1d9-f4e0-4c67-923e-2494e15c4088\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-lqmgv" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.718698 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxgwb\" (UniqueName: \"kubernetes.io/projected/898b32c5-9f21-4fba-90c5-a333f36addf2-kube-api-access-cxgwb\") pod \"placement-operator-controller-manager-5784578c99-jj4pt\" (UID: \"898b32c5-9f21-4fba-90c5-a333f36addf2\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-jj4pt" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.721555 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-l59gx"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.722413 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-l59gx" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.728055 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-9psgs" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.740804 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrnmj\" (UniqueName: \"kubernetes.io/projected/3035739a-202f-4794-bb4f-ae2342a96441-kube-api-access-jrnmj\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rwq59\" (UID: \"3035739a-202f-4794-bb4f-ae2342a96441\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.740905 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rwq59\" (UID: \"3035739a-202f-4794-bb4f-ae2342a96441\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59" Mar 21 05:04:11 crc kubenswrapper[4775]: E0321 05:04:11.741144 4775 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 05:04:11 crc kubenswrapper[4775]: E0321 05:04:11.741203 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert podName:3035739a-202f-4794-bb4f-ae2342a96441 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:12.241179823 +0000 UTC m=+1005.217643447 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-rwq59" (UID: "3035739a-202f-4794-bb4f-ae2342a96441") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.745140 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hf688" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.756821 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-767865f676-4tbvg" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.771218 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-l59gx"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.773726 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7zpk\" (UniqueName: \"kubernetes.io/projected/6eeb04ad-7251-488c-bd52-b2f14f6fb68b-kube-api-access-x7zpk\") pod \"ovn-operator-controller-manager-884679f54-ctm9h\" (UID: \"6eeb04ad-7251-488c-bd52-b2f14f6fb68b\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-ctm9h" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.774637 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxgwb\" (UniqueName: \"kubernetes.io/projected/898b32c5-9f21-4fba-90c5-a333f36addf2-kube-api-access-cxgwb\") pod \"placement-operator-controller-manager-5784578c99-jj4pt\" (UID: \"898b32c5-9f21-4fba-90c5-a333f36addf2\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-jj4pt" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.794176 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tss7r" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.787430 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrnmj\" (UniqueName: \"kubernetes.io/projected/3035739a-202f-4794-bb4f-ae2342a96441-kube-api-access-jrnmj\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rwq59\" (UID: \"3035739a-202f-4794-bb4f-ae2342a96441\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.817526 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-67scw" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.837375 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-gdrc5"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.838473 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-gdrc5" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.842534 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-tf6bp" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.843224 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thj25\" (UniqueName: \"kubernetes.io/projected/d1dbd80a-0782-4035-a263-b52a90f6ee0e-kube-api-access-thj25\") pod \"telemetry-operator-controller-manager-d6b694c5-9s82h\" (UID: \"d1dbd80a-0782-4035-a263-b52a90f6ee0e\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-9s82h" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.843335 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tdnx\" (UniqueName: \"kubernetes.io/projected/9cac78ed-6325-4649-bb05-a1518ae692e9-kube-api-access-4tdnx\") pod \"test-operator-controller-manager-5c5cb9c4d7-l59gx\" (UID: \"9cac78ed-6325-4649-bb05-a1518ae692e9\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-l59gx" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.843373 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g8sv\" (UniqueName: \"kubernetes.io/projected/5968f1d9-f4e0-4c67-923e-2494e15c4088-kube-api-access-6g8sv\") pod \"swift-operator-controller-manager-c674c5965-lqmgv\" (UID: \"5968f1d9-f4e0-4c67-923e-2494e15c4088\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-lqmgv" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.850667 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-gdrc5"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.881411 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5784578c99-jj4pt" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.890008 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g8sv\" (UniqueName: \"kubernetes.io/projected/5968f1d9-f4e0-4c67-923e-2494e15c4088-kube-api-access-6g8sv\") pod \"swift-operator-controller-manager-c674c5965-lqmgv\" (UID: \"5968f1d9-f4e0-4c67-923e-2494e15c4088\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-lqmgv" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.902900 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq"] Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.904398 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.911158 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.912201 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.912296 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-zqnnz" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.938053 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-884679f54-ctm9h" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.945101 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert\") pod \"infra-operator-controller-manager-65f65cc49c-2mgp8\" (UID: \"0a66456f-7860-4dc1-9c1c-0db69ddcc800\") " pod="openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.945176 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.945240 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thj25\" (UniqueName: \"kubernetes.io/projected/d1dbd80a-0782-4035-a263-b52a90f6ee0e-kube-api-access-thj25\") pod \"telemetry-operator-controller-manager-d6b694c5-9s82h\" (UID: \"d1dbd80a-0782-4035-a263-b52a90f6ee0e\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-9s82h" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.945316 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tdnx\" (UniqueName: \"kubernetes.io/projected/9cac78ed-6325-4649-bb05-a1518ae692e9-kube-api-access-4tdnx\") pod \"test-operator-controller-manager-5c5cb9c4d7-l59gx\" (UID: \"9cac78ed-6325-4649-bb05-a1518ae692e9\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-l59gx" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.945381 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg86m\" (UniqueName: \"kubernetes.io/projected/907f0cdf-2d87-4d09-97af-5591d061b4f6-kube-api-access-cg86m\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.945409 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:11 crc kubenswrapper[4775]: I0321 05:04:11.945429 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clvqc\" (UniqueName: \"kubernetes.io/projected/3c9f18bd-def6-45ff-a92b-25c6f40d6bb5-kube-api-access-clvqc\") pod \"watcher-operator-controller-manager-6c4d75f7f9-gdrc5\" (UID: \"3c9f18bd-def6-45ff-a92b-25c6f40d6bb5\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-gdrc5" Mar 21 05:04:11 crc kubenswrapper[4775]: E0321 05:04:11.945606 4775 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 21 05:04:11 crc kubenswrapper[4775]: E0321 05:04:11.945647 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert podName:0a66456f-7860-4dc1-9c1c-0db69ddcc800 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:12.945633088 +0000 UTC m=+1005.922096712 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert") pod "infra-operator-controller-manager-65f65cc49c-2mgp8" (UID: "0a66456f-7860-4dc1-9c1c-0db69ddcc800") : secret "infra-operator-webhook-server-cert" not found Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.022630 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-c674c5965-lqmgv" Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.042875 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq"] Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.047266 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jvglf"] Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.048065 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg86m\" (UniqueName: \"kubernetes.io/projected/907f0cdf-2d87-4d09-97af-5591d061b4f6-kube-api-access-cg86m\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.049245 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.049381 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clvqc\" (UniqueName: \"kubernetes.io/projected/3c9f18bd-def6-45ff-a92b-25c6f40d6bb5-kube-api-access-clvqc\") pod \"watcher-operator-controller-manager-6c4d75f7f9-gdrc5\" (UID: \"3c9f18bd-def6-45ff-a92b-25c6f40d6bb5\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-gdrc5" Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.049453 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jvglf" Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.049637 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:12 crc kubenswrapper[4775]: E0321 05:04:12.057290 4775 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 21 05:04:12 crc kubenswrapper[4775]: E0321 05:04:12.057619 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs podName:907f0cdf-2d87-4d09-97af-5591d061b4f6 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:12.55758417 +0000 UTC m=+1005.534047794 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs") pod "openstack-operator-controller-manager-65746ff4dc-hg4rq" (UID: "907f0cdf-2d87-4d09-97af-5591d061b4f6") : secret "webhook-server-cert" not found Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.057839 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-2ksnd" Mar 21 05:04:12 crc kubenswrapper[4775]: E0321 05:04:12.057953 4775 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 21 05:04:12 crc kubenswrapper[4775]: E0321 05:04:12.058072 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs podName:907f0cdf-2d87-4d09-97af-5591d061b4f6 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:12.558058524 +0000 UTC m=+1005.534522148 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs") pod "openstack-operator-controller-manager-65746ff4dc-hg4rq" (UID: "907f0cdf-2d87-4d09-97af-5591d061b4f6") : secret "metrics-server-cert" not found Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.057439 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thj25\" (UniqueName: \"kubernetes.io/projected/d1dbd80a-0782-4035-a263-b52a90f6ee0e-kube-api-access-thj25\") pod \"telemetry-operator-controller-manager-d6b694c5-9s82h\" (UID: \"d1dbd80a-0782-4035-a263-b52a90f6ee0e\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-9s82h" Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.077562 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-9s82h" Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.081939 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tdnx\" (UniqueName: \"kubernetes.io/projected/9cac78ed-6325-4649-bb05-a1518ae692e9-kube-api-access-4tdnx\") pod \"test-operator-controller-manager-5c5cb9c4d7-l59gx\" (UID: \"9cac78ed-6325-4649-bb05-a1518ae692e9\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-l59gx" Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.093952 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-l59gx" Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.125148 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clvqc\" (UniqueName: \"kubernetes.io/projected/3c9f18bd-def6-45ff-a92b-25c6f40d6bb5-kube-api-access-clvqc\") pod \"watcher-operator-controller-manager-6c4d75f7f9-gdrc5\" (UID: \"3c9f18bd-def6-45ff-a92b-25c6f40d6bb5\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-gdrc5" Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.141774 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg86m\" (UniqueName: \"kubernetes.io/projected/907f0cdf-2d87-4d09-97af-5591d061b4f6-kube-api-access-cg86m\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.184834 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jvglf"] Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.204266 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-gdrc5" Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.211052 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6nhhl"] Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.223198 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-dn22m"] Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.238541 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-lxvtw"] Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.248948 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-twhxx"] Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.259870 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rwq59\" (UID: \"3035739a-202f-4794-bb4f-ae2342a96441\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59" Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.259976 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdx89\" (UniqueName: \"kubernetes.io/projected/95a44b12-e027-400d-b257-99f2012251d8-kube-api-access-kdx89\") pod \"rabbitmq-cluster-operator-manager-668c99d594-jvglf\" (UID: \"95a44b12-e027-400d-b257-99f2012251d8\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jvglf" Mar 21 05:04:12 crc kubenswrapper[4775]: E0321 05:04:12.260173 4775 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 05:04:12 crc kubenswrapper[4775]: E0321 05:04:12.260213 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert podName:3035739a-202f-4794-bb4f-ae2342a96441 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:13.260199123 +0000 UTC m=+1006.236662747 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-rwq59" (UID: "3035739a-202f-4794-bb4f-ae2342a96441") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.363438 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdx89\" (UniqueName: \"kubernetes.io/projected/95a44b12-e027-400d-b257-99f2012251d8-kube-api-access-kdx89\") pod \"rabbitmq-cluster-operator-manager-668c99d594-jvglf\" (UID: \"95a44b12-e027-400d-b257-99f2012251d8\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jvglf" Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.386302 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdx89\" (UniqueName: \"kubernetes.io/projected/95a44b12-e027-400d-b257-99f2012251d8-kube-api-access-kdx89\") pod \"rabbitmq-cluster-operator-manager-668c99d594-jvglf\" (UID: \"95a44b12-e027-400d-b257-99f2012251d8\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jvglf" Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.431215 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-lzvsq"] Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.455527 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-m4gqz"] Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.523812 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jvglf" Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.565856 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:12 crc kubenswrapper[4775]: E0321 05:04:12.565967 4775 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.565993 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:12 crc kubenswrapper[4775]: E0321 05:04:12.566034 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs podName:907f0cdf-2d87-4d09-97af-5591d061b4f6 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:13.56601363 +0000 UTC m=+1006.542477254 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs") pod "openstack-operator-controller-manager-65746ff4dc-hg4rq" (UID: "907f0cdf-2d87-4d09-97af-5591d061b4f6") : secret "webhook-server-cert" not found Mar 21 05:04:12 crc kubenswrapper[4775]: E0321 05:04:12.566100 4775 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 21 05:04:12 crc kubenswrapper[4775]: E0321 05:04:12.566151 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs podName:907f0cdf-2d87-4d09-97af-5591d061b4f6 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:13.566139723 +0000 UTC m=+1006.542603347 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs") pod "openstack-operator-controller-manager-65746ff4dc-hg4rq" (UID: "907f0cdf-2d87-4d09-97af-5591d061b4f6") : secret "metrics-server-cert" not found Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.574193 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-7bdbc"] Mar 21 05:04:12 crc kubenswrapper[4775]: W0321 05:04:12.611806 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01ed348d_8a8a_4717_ba0d_1944b3f1c081.slice/crio-370ea74a650349df6fa12d38039e6e6c3bd1b1f3b383cb055129fd355e72f1c6 WatchSource:0}: Error finding container 370ea74a650349df6fa12d38039e6e6c3bd1b1f3b383cb055129fd355e72f1c6: Status 404 returned error can't find the container with id 370ea74a650349df6fa12d38039e6e6c3bd1b1f3b383cb055129fd355e72f1c6 Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.902386 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-9g8pv"] Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.909555 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-c7pjw"] Mar 21 05:04:12 crc kubenswrapper[4775]: W0321 05:04:12.922292 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf00c8c4b_874f_45ec_8a1a_e0834b3fc252.slice/crio-9952a05dffed9417a338c1437660f1e8c0b0b0d0d8381950453c223b515205a5 WatchSource:0}: Error finding container 9952a05dffed9417a338c1437660f1e8c0b0b0d0d8381950453c223b515205a5: Status 404 returned error can't find the container with id 9952a05dffed9417a338c1437660f1e8c0b0b0d0d8381950453c223b515205a5 Mar 21 05:04:12 crc kubenswrapper[4775]: W0321 05:04:12.924395 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49a3c6c7_9e86_495d_8d1a_486d6bfbbbdd.slice/crio-0e5d6af419e11ff092281a5770d12a7c04f50f7843675dbc29cbb9a75a9d66fb WatchSource:0}: Error finding container 0e5d6af419e11ff092281a5770d12a7c04f50f7843675dbc29cbb9a75a9d66fb: Status 404 returned error can't find the container with id 0e5d6af419e11ff092281a5770d12a7c04f50f7843675dbc29cbb9a75a9d66fb Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.925532 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-jq88h"] Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.929687 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-twhxx" event={"ID":"94e1507b-be6c-4ecf-99e5-2bdcd2cc0cef","Type":"ContainerStarted","Data":"9463c706f5f47e68a62ef68715be3e8c7f5a1b32aadc69665b5db9b96a7ce7bd"} Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.934150 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-m4gqz" event={"ID":"0e83601c-758c-4f12-b745-bb68b0c4904f","Type":"ContainerStarted","Data":"7c70d5135af72bc479363a2adcde7e746fa83857807b8d04642a6365ca7ae8f5"} Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.936185 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-lzvsq" event={"ID":"80932361-6406-48dd-9e4b-4e9c27813f68","Type":"ContainerStarted","Data":"d9a8ee0cbb4e1d1ae9a8e8706a43da215a4d62cb7e6828d519feacdd63413f2e"} Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.937987 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-dn22m" event={"ID":"9fe71acc-7d35-4d4b-ac69-e193d3f39028","Type":"ContainerStarted","Data":"f192a51da0996ede725942f75554886d6952f4b898a6ee74b81f602102e264c8"} Mar 21 05:04:12 crc kubenswrapper[4775]: W0321 05:04:12.938262 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20c78c73_daf5_481e_a4ac_62de73b5969e.slice/crio-93ad1f6e6168188143481da9a75415dc36e74d1bf96d6519927b1c6f285c4852 WatchSource:0}: Error finding container 93ad1f6e6168188143481da9a75415dc36e74d1bf96d6519927b1c6f285c4852: Status 404 returned error can't find the container with id 93ad1f6e6168188143481da9a75415dc36e74d1bf96d6519927b1c6f285c4852 Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.943866 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-lxvtw" event={"ID":"87dcea67-7f65-46a6-996b-3985bf1b5171","Type":"ContainerStarted","Data":"f8bf0412db315132c2c4a3b4becf65c579aea5c697c3bbf4a646c390dab460a7"} Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.946691 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6nhhl" podUID="694f1c66-ce56-4854-9f4c-02a42e16046d" containerName="registry-server" containerID="cri-o://bfa25ad839c3791aa796fdbb1ca916c818ffc84d5a2bd52f3d19c4a9886bd0e1" gracePeriod=2 Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.946773 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-7bdbc" event={"ID":"01ed348d-8a8a-4717-ba0d-1944b3f1c081","Type":"ContainerStarted","Data":"370ea74a650349df6fa12d38039e6e6c3bd1b1f3b383cb055129fd355e72f1c6"} Mar 21 05:04:12 crc kubenswrapper[4775]: I0321 05:04:12.971873 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert\") pod \"infra-operator-controller-manager-65f65cc49c-2mgp8\" (UID: \"0a66456f-7860-4dc1-9c1c-0db69ddcc800\") " pod="openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8" Mar 21 05:04:12 crc kubenswrapper[4775]: E0321 05:04:12.972100 4775 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 21 05:04:12 crc kubenswrapper[4775]: E0321 05:04:12.972179 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert podName:0a66456f-7860-4dc1-9c1c-0db69ddcc800 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:14.97215903 +0000 UTC m=+1007.948622654 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert") pod "infra-operator-controller-manager-65f65cc49c-2mgp8" (UID: "0a66456f-7860-4dc1-9c1c-0db69ddcc800") : secret "infra-operator-webhook-server-cert" not found Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.229576 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-67scw"] Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.235486 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-9s82h"] Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.256739 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-lqmgv"] Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.262755 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-jj4pt"] Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.277460 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rwq59\" (UID: \"3035739a-202f-4794-bb4f-ae2342a96441\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59" Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.277938 4775 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.277994 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert podName:3035739a-202f-4794-bb4f-ae2342a96441 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:15.277976727 +0000 UTC m=+1008.254440351 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-rwq59" (UID: "3035739a-202f-4794-bb4f-ae2342a96441") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.278402 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-tss7r"] Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.286194 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-ctm9h"] Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.299312 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hf688"] Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.305224 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-l59gx"] Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.312351 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-4tbvg"] Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.321211 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-gdrc5"] Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.324358 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jvglf"] Mar 21 05:04:13 crc kubenswrapper[4775]: W0321 05:04:13.350896 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c9f18bd_def6_45ff_a92b_25c6f40d6bb5.slice/crio-ad1a20a2a60b39e31bf0dc14b2a2a50fe4d299b742bce9bbf2eed2a2b544d612 WatchSource:0}: Error finding container ad1a20a2a60b39e31bf0dc14b2a2a50fe4d299b742bce9bbf2eed2a2b544d612: Status 404 returned error can't find the container with id ad1a20a2a60b39e31bf0dc14b2a2a50fe4d299b742bce9bbf2eed2a2b544d612 Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.351485 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kdx89,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-jvglf_openstack-operators(95a44b12-e027-400d-b257-99f2012251d8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.351603 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:bef93f71d3b42a72d8b96c69bdb4db4b8bd797c5093a0a719443d7a5c9aaab55,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x7zpk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-884679f54-ctm9h_openstack-operators(6eeb04ad-7251-488c-bd52-b2f14f6fb68b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.351629 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:526f9d4965431e1a5e4f8c3224bcee3f636a3108a5e0767296a994c2a517404a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8zv76,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-767865f676-4tbvg_openstack-operators(9bfe7d25-53ea-484d-a481-0ea04ee2b8a8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.351747 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:6e7552996253fc66667eaa3eb0e11b4e97145efa2ae577155ceabf8e9913ddc1,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-84jdm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-67ccfc9778-hf688_openstack-operators(ec38d53e-6fe4-41b5-8548-e49fadd9d6bf): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.353158 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-767865f676-4tbvg" podUID="9bfe7d25-53ea-484d-a481-0ea04ee2b8a8" Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.353203 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hf688" podUID="ec38d53e-6fe4-41b5-8548-e49fadd9d6bf" Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.353225 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-884679f54-ctm9h" podUID="6eeb04ad-7251-488c-bd52-b2f14f6fb68b" Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.353244 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jvglf" podUID="95a44b12-e027-400d-b257-99f2012251d8" Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.368329 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-clvqc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6c4d75f7f9-gdrc5_openstack-operators(3c9f18bd-def6-45ff-a92b-25c6f40d6bb5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.369498 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-gdrc5" podUID="3c9f18bd-def6-45ff-a92b-25c6f40d6bb5" Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.581022 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.581127 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.581223 4775 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.581281 4775 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.581350 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs podName:907f0cdf-2d87-4d09-97af-5591d061b4f6 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:15.581325084 +0000 UTC m=+1008.557788708 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs") pod "openstack-operator-controller-manager-65746ff4dc-hg4rq" (UID: "907f0cdf-2d87-4d09-97af-5591d061b4f6") : secret "metrics-server-cert" not found Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.581370 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs podName:907f0cdf-2d87-4d09-97af-5591d061b4f6 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:15.581362275 +0000 UTC m=+1008.557825899 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs") pod "openstack-operator-controller-manager-65746ff4dc-hg4rq" (UID: "907f0cdf-2d87-4d09-97af-5591d061b4f6") : secret "webhook-server-cert" not found Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.955176 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-c674c5965-lqmgv" event={"ID":"5968f1d9-f4e0-4c67-923e-2494e15c4088","Type":"ContainerStarted","Data":"c22a3b421b20aebd6844914d88ac1b1eca165d2c5ff23665d7ffe3ed50be4b1e"} Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.958272 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tss7r" event={"ID":"02b6af47-2c06-480b-a838-2d742efa1045","Type":"ContainerStarted","Data":"7906999cc16e82fcb0dcd2e5020ffe06fa838cf9857fb99542fdcc09100516ff"} Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.959260 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-67scw" event={"ID":"745c79b1-1bcf-4c0f-82ee-a26cbba46d48","Type":"ContainerStarted","Data":"95e47c4facffa5845ce3352cfbc986a6018f2c261d3b23fb69c8a344f68ab3c9"} Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.961780 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hf688" event={"ID":"ec38d53e-6fe4-41b5-8548-e49fadd9d6bf","Type":"ContainerStarted","Data":"543341eb244d971192be7aebcdafb74132e2b5df8f2c5068cf04dab2caab621a"} Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.964249 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:6e7552996253fc66667eaa3eb0e11b4e97145efa2ae577155ceabf8e9913ddc1\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hf688" podUID="ec38d53e-6fe4-41b5-8548-e49fadd9d6bf" Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.966900 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5784578c99-jj4pt" event={"ID":"898b32c5-9f21-4fba-90c5-a333f36addf2","Type":"ContainerStarted","Data":"dc24878786da3f3dbe3c3011dfcb502d4c5a8dfec748074f92c806f91a7029a0"} Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.969138 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-9s82h" event={"ID":"d1dbd80a-0782-4035-a263-b52a90f6ee0e","Type":"ContainerStarted","Data":"018b768dcada9e8a05496b77c87fd5746e1d3d786081d4cceed611a15644df26"} Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.971030 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-9g8pv" event={"ID":"f00c8c4b-874f-45ec-8a1a-e0834b3fc252","Type":"ContainerStarted","Data":"9952a05dffed9417a338c1437660f1e8c0b0b0d0d8381950453c223b515205a5"} Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.972145 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jvglf" event={"ID":"95a44b12-e027-400d-b257-99f2012251d8","Type":"ContainerStarted","Data":"806e83334c792a8b0305a7f67d211b4f3a82d2f33b6b96148798e5a534611c0d"} Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.973545 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jvglf" podUID="95a44b12-e027-400d-b257-99f2012251d8" Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.974607 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-55f864c847-jq88h" event={"ID":"20c78c73-daf5-481e-a4ac-62de73b5969e","Type":"ContainerStarted","Data":"93ad1f6e6168188143481da9a75415dc36e74d1bf96d6519927b1c6f285c4852"} Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.975776 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c7pjw" event={"ID":"49a3c6c7-9e86-495d-8d1a-486d6bfbbbdd","Type":"ContainerStarted","Data":"0e5d6af419e11ff092281a5770d12a7c04f50f7843675dbc29cbb9a75a9d66fb"} Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.976776 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-l59gx" event={"ID":"9cac78ed-6325-4649-bb05-a1518ae692e9","Type":"ContainerStarted","Data":"d33f5adf3ebacaf877df27a18755fffa61d1f5869b445de66666e8c34353580e"} Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.977622 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-767865f676-4tbvg" event={"ID":"9bfe7d25-53ea-484d-a481-0ea04ee2b8a8","Type":"ContainerStarted","Data":"e28c181cd4648baa68bcef5fa441b083107a741b5c2a8f3ec951c960aabffcf3"} Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.979406 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:526f9d4965431e1a5e4f8c3224bcee3f636a3108a5e0767296a994c2a517404a\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-767865f676-4tbvg" podUID="9bfe7d25-53ea-484d-a481-0ea04ee2b8a8" Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.981192 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-gdrc5" event={"ID":"3c9f18bd-def6-45ff-a92b-25c6f40d6bb5","Type":"ContainerStarted","Data":"ad1a20a2a60b39e31bf0dc14b2a2a50fe4d299b742bce9bbf2eed2a2b544d612"} Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.982325 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-gdrc5" podUID="3c9f18bd-def6-45ff-a92b-25c6f40d6bb5" Mar 21 05:04:13 crc kubenswrapper[4775]: I0321 05:04:13.982876 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-884679f54-ctm9h" event={"ID":"6eeb04ad-7251-488c-bd52-b2f14f6fb68b","Type":"ContainerStarted","Data":"148ea4025f8ef65bc1a19cec65c2ba6919e2bb2969fbd9638161e6500d07dc60"} Mar 21 05:04:13 crc kubenswrapper[4775]: E0321 05:04:13.984448 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:bef93f71d3b42a72d8b96c69bdb4db4b8bd797c5093a0a719443d7a5c9aaab55\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-884679f54-ctm9h" podUID="6eeb04ad-7251-488c-bd52-b2f14f6fb68b" Mar 21 05:04:15 crc kubenswrapper[4775]: I0321 05:04:15.003864 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert\") pod \"infra-operator-controller-manager-65f65cc49c-2mgp8\" (UID: \"0a66456f-7860-4dc1-9c1c-0db69ddcc800\") " pod="openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8" Mar 21 05:04:15 crc kubenswrapper[4775]: E0321 05:04:15.004025 4775 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 21 05:04:15 crc kubenswrapper[4775]: E0321 05:04:15.004352 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert podName:0a66456f-7860-4dc1-9c1c-0db69ddcc800 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:19.004327253 +0000 UTC m=+1011.980790867 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert") pod "infra-operator-controller-manager-65f65cc49c-2mgp8" (UID: "0a66456f-7860-4dc1-9c1c-0db69ddcc800") : secret "infra-operator-webhook-server-cert" not found Mar 21 05:04:15 crc kubenswrapper[4775]: I0321 05:04:15.018408 4775 generic.go:334] "Generic (PLEG): container finished" podID="694f1c66-ce56-4854-9f4c-02a42e16046d" containerID="bfa25ad839c3791aa796fdbb1ca916c818ffc84d5a2bd52f3d19c4a9886bd0e1" exitCode=0 Mar 21 05:04:15 crc kubenswrapper[4775]: I0321 05:04:15.018489 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nhhl" event={"ID":"694f1c66-ce56-4854-9f4c-02a42e16046d","Type":"ContainerDied","Data":"bfa25ad839c3791aa796fdbb1ca916c818ffc84d5a2bd52f3d19c4a9886bd0e1"} Mar 21 05:04:15 crc kubenswrapper[4775]: E0321 05:04:15.023991 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:6e7552996253fc66667eaa3eb0e11b4e97145efa2ae577155ceabf8e9913ddc1\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hf688" podUID="ec38d53e-6fe4-41b5-8548-e49fadd9d6bf" Mar 21 05:04:15 crc kubenswrapper[4775]: E0321 05:04:15.024059 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-gdrc5" podUID="3c9f18bd-def6-45ff-a92b-25c6f40d6bb5" Mar 21 05:04:15 crc kubenswrapper[4775]: E0321 05:04:15.024284 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:bef93f71d3b42a72d8b96c69bdb4db4b8bd797c5093a0a719443d7a5c9aaab55\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-884679f54-ctm9h" podUID="6eeb04ad-7251-488c-bd52-b2f14f6fb68b" Mar 21 05:04:15 crc kubenswrapper[4775]: E0321 05:04:15.024334 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:526f9d4965431e1a5e4f8c3224bcee3f636a3108a5e0767296a994c2a517404a\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-767865f676-4tbvg" podUID="9bfe7d25-53ea-484d-a481-0ea04ee2b8a8" Mar 21 05:04:15 crc kubenswrapper[4775]: E0321 05:04:15.026340 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jvglf" podUID="95a44b12-e027-400d-b257-99f2012251d8" Mar 21 05:04:15 crc kubenswrapper[4775]: I0321 05:04:15.309887 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rwq59\" (UID: \"3035739a-202f-4794-bb4f-ae2342a96441\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59" Mar 21 05:04:15 crc kubenswrapper[4775]: E0321 05:04:15.310130 4775 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 05:04:15 crc kubenswrapper[4775]: E0321 05:04:15.310188 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert podName:3035739a-202f-4794-bb4f-ae2342a96441 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:19.31017322 +0000 UTC m=+1012.286636844 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-rwq59" (UID: "3035739a-202f-4794-bb4f-ae2342a96441") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 05:04:15 crc kubenswrapper[4775]: I0321 05:04:15.616362 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:15 crc kubenswrapper[4775]: I0321 05:04:15.616481 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:15 crc kubenswrapper[4775]: E0321 05:04:15.616531 4775 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 21 05:04:15 crc kubenswrapper[4775]: E0321 05:04:15.616677 4775 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 21 05:04:15 crc kubenswrapper[4775]: E0321 05:04:15.616713 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs podName:907f0cdf-2d87-4d09-97af-5591d061b4f6 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:19.616691197 +0000 UTC m=+1012.593154821 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs") pod "openstack-operator-controller-manager-65746ff4dc-hg4rq" (UID: "907f0cdf-2d87-4d09-97af-5591d061b4f6") : secret "metrics-server-cert" not found Mar 21 05:04:15 crc kubenswrapper[4775]: E0321 05:04:15.616734 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs podName:907f0cdf-2d87-4d09-97af-5591d061b4f6 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:19.616726068 +0000 UTC m=+1012.593189692 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs") pod "openstack-operator-controller-manager-65746ff4dc-hg4rq" (UID: "907f0cdf-2d87-4d09-97af-5591d061b4f6") : secret "webhook-server-cert" not found Mar 21 05:04:15 crc kubenswrapper[4775]: I0321 05:04:15.804424 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nhhl" Mar 21 05:04:15 crc kubenswrapper[4775]: I0321 05:04:15.919803 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/694f1c66-ce56-4854-9f4c-02a42e16046d-utilities\") pod \"694f1c66-ce56-4854-9f4c-02a42e16046d\" (UID: \"694f1c66-ce56-4854-9f4c-02a42e16046d\") " Mar 21 05:04:15 crc kubenswrapper[4775]: I0321 05:04:15.919940 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xg6ln\" (UniqueName: \"kubernetes.io/projected/694f1c66-ce56-4854-9f4c-02a42e16046d-kube-api-access-xg6ln\") pod \"694f1c66-ce56-4854-9f4c-02a42e16046d\" (UID: \"694f1c66-ce56-4854-9f4c-02a42e16046d\") " Mar 21 05:04:15 crc kubenswrapper[4775]: I0321 05:04:15.919964 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/694f1c66-ce56-4854-9f4c-02a42e16046d-catalog-content\") pod \"694f1c66-ce56-4854-9f4c-02a42e16046d\" (UID: \"694f1c66-ce56-4854-9f4c-02a42e16046d\") " Mar 21 05:04:15 crc kubenswrapper[4775]: I0321 05:04:15.920728 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/694f1c66-ce56-4854-9f4c-02a42e16046d-utilities" (OuterVolumeSpecName: "utilities") pod "694f1c66-ce56-4854-9f4c-02a42e16046d" (UID: "694f1c66-ce56-4854-9f4c-02a42e16046d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:04:15 crc kubenswrapper[4775]: I0321 05:04:15.931180 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/694f1c66-ce56-4854-9f4c-02a42e16046d-kube-api-access-xg6ln" (OuterVolumeSpecName: "kube-api-access-xg6ln") pod "694f1c66-ce56-4854-9f4c-02a42e16046d" (UID: "694f1c66-ce56-4854-9f4c-02a42e16046d"). InnerVolumeSpecName "kube-api-access-xg6ln". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:04:15 crc kubenswrapper[4775]: I0321 05:04:15.987401 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/694f1c66-ce56-4854-9f4c-02a42e16046d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "694f1c66-ce56-4854-9f4c-02a42e16046d" (UID: "694f1c66-ce56-4854-9f4c-02a42e16046d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:04:16 crc kubenswrapper[4775]: I0321 05:04:16.021670 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/694f1c66-ce56-4854-9f4c-02a42e16046d-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:04:16 crc kubenswrapper[4775]: I0321 05:04:16.021706 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xg6ln\" (UniqueName: \"kubernetes.io/projected/694f1c66-ce56-4854-9f4c-02a42e16046d-kube-api-access-xg6ln\") on node \"crc\" DevicePath \"\"" Mar 21 05:04:16 crc kubenswrapper[4775]: I0321 05:04:16.021717 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/694f1c66-ce56-4854-9f4c-02a42e16046d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:04:16 crc kubenswrapper[4775]: I0321 05:04:16.032433 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nhhl" event={"ID":"694f1c66-ce56-4854-9f4c-02a42e16046d","Type":"ContainerDied","Data":"3f6120ae96f0b10b5bf76bd3efa458fabb2c0680a195b83d1f67570f61231bb6"} Mar 21 05:04:16 crc kubenswrapper[4775]: I0321 05:04:16.032482 4775 scope.go:117] "RemoveContainer" containerID="bfa25ad839c3791aa796fdbb1ca916c818ffc84d5a2bd52f3d19c4a9886bd0e1" Mar 21 05:04:16 crc kubenswrapper[4775]: I0321 05:04:16.032498 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nhhl" Mar 21 05:04:16 crc kubenswrapper[4775]: I0321 05:04:16.067174 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6nhhl"] Mar 21 05:04:16 crc kubenswrapper[4775]: I0321 05:04:16.071539 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6nhhl"] Mar 21 05:04:17 crc kubenswrapper[4775]: I0321 05:04:17.674009 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="694f1c66-ce56-4854-9f4c-02a42e16046d" path="/var/lib/kubelet/pods/694f1c66-ce56-4854-9f4c-02a42e16046d/volumes" Mar 21 05:04:19 crc kubenswrapper[4775]: I0321 05:04:19.069971 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert\") pod \"infra-operator-controller-manager-65f65cc49c-2mgp8\" (UID: \"0a66456f-7860-4dc1-9c1c-0db69ddcc800\") " pod="openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8" Mar 21 05:04:19 crc kubenswrapper[4775]: E0321 05:04:19.070269 4775 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 21 05:04:19 crc kubenswrapper[4775]: E0321 05:04:19.070845 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert podName:0a66456f-7860-4dc1-9c1c-0db69ddcc800 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:27.070820738 +0000 UTC m=+1020.047284362 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert") pod "infra-operator-controller-manager-65f65cc49c-2mgp8" (UID: "0a66456f-7860-4dc1-9c1c-0db69ddcc800") : secret "infra-operator-webhook-server-cert" not found Mar 21 05:04:19 crc kubenswrapper[4775]: I0321 05:04:19.374776 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rwq59\" (UID: \"3035739a-202f-4794-bb4f-ae2342a96441\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59" Mar 21 05:04:19 crc kubenswrapper[4775]: E0321 05:04:19.375002 4775 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 05:04:19 crc kubenswrapper[4775]: E0321 05:04:19.375131 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert podName:3035739a-202f-4794-bb4f-ae2342a96441 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:27.375091901 +0000 UTC m=+1020.351555535 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-rwq59" (UID: "3035739a-202f-4794-bb4f-ae2342a96441") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 05:04:19 crc kubenswrapper[4775]: I0321 05:04:19.679842 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:19 crc kubenswrapper[4775]: I0321 05:04:19.680004 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:19 crc kubenswrapper[4775]: E0321 05:04:19.680180 4775 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 21 05:04:19 crc kubenswrapper[4775]: E0321 05:04:19.680218 4775 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 21 05:04:19 crc kubenswrapper[4775]: E0321 05:04:19.680303 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs podName:907f0cdf-2d87-4d09-97af-5591d061b4f6 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:27.68027678 +0000 UTC m=+1020.656740404 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs") pod "openstack-operator-controller-manager-65746ff4dc-hg4rq" (UID: "907f0cdf-2d87-4d09-97af-5591d061b4f6") : secret "metrics-server-cert" not found Mar 21 05:04:19 crc kubenswrapper[4775]: E0321 05:04:19.680329 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs podName:907f0cdf-2d87-4d09-97af-5591d061b4f6 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:27.680317782 +0000 UTC m=+1020.656781406 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs") pod "openstack-operator-controller-manager-65746ff4dc-hg4rq" (UID: "907f0cdf-2d87-4d09-97af-5591d061b4f6") : secret "webhook-server-cert" not found Mar 21 05:04:27 crc kubenswrapper[4775]: I0321 05:04:27.095002 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert\") pod \"infra-operator-controller-manager-65f65cc49c-2mgp8\" (UID: \"0a66456f-7860-4dc1-9c1c-0db69ddcc800\") " pod="openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8" Mar 21 05:04:27 crc kubenswrapper[4775]: E0321 05:04:27.095190 4775 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 21 05:04:27 crc kubenswrapper[4775]: E0321 05:04:27.095684 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert podName:0a66456f-7860-4dc1-9c1c-0db69ddcc800 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:43.095668035 +0000 UTC m=+1036.072131659 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert") pod "infra-operator-controller-manager-65f65cc49c-2mgp8" (UID: "0a66456f-7860-4dc1-9c1c-0db69ddcc800") : secret "infra-operator-webhook-server-cert" not found Mar 21 05:04:27 crc kubenswrapper[4775]: I0321 05:04:27.289805 4775 scope.go:117] "RemoveContainer" containerID="428647d2139b15f3ab85f6734bc25c0214e38d663b4a1c7cb9b69a834fdbe970" Mar 21 05:04:27 crc kubenswrapper[4775]: I0321 05:04:27.400441 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rwq59\" (UID: \"3035739a-202f-4794-bb4f-ae2342a96441\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59" Mar 21 05:04:27 crc kubenswrapper[4775]: E0321 05:04:27.400707 4775 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 05:04:27 crc kubenswrapper[4775]: E0321 05:04:27.400900 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert podName:3035739a-202f-4794-bb4f-ae2342a96441 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:43.400878605 +0000 UTC m=+1036.377342239 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-rwq59" (UID: "3035739a-202f-4794-bb4f-ae2342a96441") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 05:04:27 crc kubenswrapper[4775]: I0321 05:04:27.706651 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:27 crc kubenswrapper[4775]: I0321 05:04:27.706782 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:27 crc kubenswrapper[4775]: E0321 05:04:27.706938 4775 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 21 05:04:27 crc kubenswrapper[4775]: E0321 05:04:27.707156 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs podName:907f0cdf-2d87-4d09-97af-5591d061b4f6 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:43.707134294 +0000 UTC m=+1036.683597938 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs") pod "openstack-operator-controller-manager-65746ff4dc-hg4rq" (UID: "907f0cdf-2d87-4d09-97af-5591d061b4f6") : secret "webhook-server-cert" not found Mar 21 05:04:27 crc kubenswrapper[4775]: E0321 05:04:27.707353 4775 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 21 05:04:27 crc kubenswrapper[4775]: E0321 05:04:27.707454 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs podName:907f0cdf-2d87-4d09-97af-5591d061b4f6 nodeName:}" failed. No retries permitted until 2026-03-21 05:04:43.707396922 +0000 UTC m=+1036.683860626 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs") pod "openstack-operator-controller-manager-65746ff4dc-hg4rq" (UID: "907f0cdf-2d87-4d09-97af-5591d061b4f6") : secret "metrics-server-cert" not found Mar 21 05:04:27 crc kubenswrapper[4775]: E0321 05:04:27.877073 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56" Mar 21 05:04:27 crc kubenswrapper[4775]: E0321 05:04:27.877597 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-66ql6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-768b96df4c-c7pjw_openstack-operators(49a3c6c7-9e86-495d-8d1a-486d6bfbbbdd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 05:04:27 crc kubenswrapper[4775]: E0321 05:04:27.879336 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c7pjw" podUID="49a3c6c7-9e86-495d-8d1a-486d6bfbbbdd" Mar 21 05:04:28 crc kubenswrapper[4775]: E0321 05:04:28.127348 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c7pjw" podUID="49a3c6c7-9e86-495d-8d1a-486d6bfbbbdd" Mar 21 05:04:28 crc kubenswrapper[4775]: I0321 05:04:28.497386 4775 scope.go:117] "RemoveContainer" containerID="5780c596a9174ce034338976560a433a9f2ac7cb3aacc03fbb37bc5801d6945f" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.163361 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-c674c5965-lqmgv" event={"ID":"5968f1d9-f4e0-4c67-923e-2494e15c4088","Type":"ContainerStarted","Data":"70696836ff0e4143f237059b34fd8d37fdd8ca4839c9b8f25d96699ea9ab6f17"} Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.164614 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-c674c5965-lqmgv" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.167490 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-dn22m" event={"ID":"9fe71acc-7d35-4d4b-ac69-e193d3f39028","Type":"ContainerStarted","Data":"d86b0eb4fc5ac0e1148a2e480f6a5150f677dc209391e3444eab379696449830"} Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.167617 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-dn22m" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.174272 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-twhxx" event={"ID":"94e1507b-be6c-4ecf-99e5-2bdcd2cc0cef","Type":"ContainerStarted","Data":"ee7de642c14e7c121b797dcc7b48f25edaca96fc418edb9a96735178a2faf5cb"} Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.174700 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-twhxx" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.183915 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5784578c99-jj4pt" event={"ID":"898b32c5-9f21-4fba-90c5-a333f36addf2","Type":"ContainerStarted","Data":"1a40b0f2ff98417dd04737fdafd26d4727fd546a6a1a9fe86114fe4c03484219"} Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.184631 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5784578c99-jj4pt" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.198402 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-m4gqz" event={"ID":"0e83601c-758c-4f12-b745-bb68b0c4904f","Type":"ContainerStarted","Data":"4db7d98f3b32e2b010f8cbc92ce1f2cd269ba478ee31833e3ae2d5dd0db7771c"} Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.199145 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-m4gqz" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.218970 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-9g8pv" event={"ID":"f00c8c4b-874f-45ec-8a1a-e0834b3fc252","Type":"ContainerStarted","Data":"95e910eb8aaa7fe737248a37785fec97e0489fab14a673d21fae213a2802dd26"} Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.219843 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-9g8pv" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.222431 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-dn22m" podStartSLOduration=2.959640905 podStartE2EDuration="19.222418288s" podCreationTimestamp="2026-03-21 05:04:10 +0000 UTC" firstStartedPulling="2026-03-21 05:04:12.212263484 +0000 UTC m=+1005.188727098" lastFinishedPulling="2026-03-21 05:04:28.475040857 +0000 UTC m=+1021.451504481" observedRunningTime="2026-03-21 05:04:29.22036637 +0000 UTC m=+1022.196829994" watchObservedRunningTime="2026-03-21 05:04:29.222418288 +0000 UTC m=+1022.198881912" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.235518 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-c674c5965-lqmgv" podStartSLOduration=3.076224519 podStartE2EDuration="18.235501189s" podCreationTimestamp="2026-03-21 05:04:11 +0000 UTC" firstStartedPulling="2026-03-21 05:04:13.335540198 +0000 UTC m=+1006.312003812" lastFinishedPulling="2026-03-21 05:04:28.494816858 +0000 UTC m=+1021.471280482" observedRunningTime="2026-03-21 05:04:29.1928296 +0000 UTC m=+1022.169293224" watchObservedRunningTime="2026-03-21 05:04:29.235501189 +0000 UTC m=+1022.211964813" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.248363 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5784578c99-jj4pt" podStartSLOduration=3.081477979 podStartE2EDuration="18.248347063s" podCreationTimestamp="2026-03-21 05:04:11 +0000 UTC" firstStartedPulling="2026-03-21 05:04:13.330581438 +0000 UTC m=+1006.307045062" lastFinishedPulling="2026-03-21 05:04:28.497450512 +0000 UTC m=+1021.473914146" observedRunningTime="2026-03-21 05:04:29.24646943 +0000 UTC m=+1022.222933054" watchObservedRunningTime="2026-03-21 05:04:29.248347063 +0000 UTC m=+1022.224810687" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.261956 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-67scw" event={"ID":"745c79b1-1bcf-4c0f-82ee-a26cbba46d48","Type":"ContainerStarted","Data":"d6dacc08d0b5fc70453c979724b352358b60f46a379573af9e32ec3ebcfecca5"} Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.262107 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-67scw" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.277237 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-55f864c847-jq88h" event={"ID":"20c78c73-daf5-481e-a4ac-62de73b5969e","Type":"ContainerStarted","Data":"f38e961c3dacd609f00cdde887590e2509ab4b7c8f5532c2b8f9b831d7b14589"} Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.278450 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-55f864c847-jq88h" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.281695 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-twhxx" podStartSLOduration=5.52008238 podStartE2EDuration="19.281679648s" podCreationTimestamp="2026-03-21 05:04:10 +0000 UTC" firstStartedPulling="2026-03-21 05:04:12.232270521 +0000 UTC m=+1005.208734145" lastFinishedPulling="2026-03-21 05:04:25.993867769 +0000 UTC m=+1018.970331413" observedRunningTime="2026-03-21 05:04:29.277287513 +0000 UTC m=+1022.253751137" watchObservedRunningTime="2026-03-21 05:04:29.281679648 +0000 UTC m=+1022.258143272" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.290725 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-7bdbc" event={"ID":"01ed348d-8a8a-4717-ba0d-1944b3f1c081","Type":"ContainerStarted","Data":"ef3f8a1625951c2579ed0041d2865d7f33f87f1596a3a6638ec0d1956bd20e81"} Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.290888 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-7bdbc" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.292528 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-lzvsq" event={"ID":"80932361-6406-48dd-9e4b-4e9c27813f68","Type":"ContainerStarted","Data":"0b58b73bdee75df59b7a689dec683772bc72f71c86539f887e29004c32c54513"} Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.293060 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-lzvsq" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.298987 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-lxvtw" event={"ID":"87dcea67-7f65-46a6-996b-3985bf1b5171","Type":"ContainerStarted","Data":"4cf5f0d88106b727b54d14ce775a20511fe8a7ddffdd0c044cf055e64587d28c"} Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.299675 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-lxvtw" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.301702 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-l59gx" event={"ID":"9cac78ed-6325-4649-bb05-a1518ae692e9","Type":"ContainerStarted","Data":"6e15552f7b63ced1a696101c769a57ef850ee06be84603155bd1e7b8d4a0ab75"} Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.302726 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-l59gx" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.315611 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-9s82h" event={"ID":"d1dbd80a-0782-4035-a263-b52a90f6ee0e","Type":"ContainerStarted","Data":"9b0cd4027cc0b41c55b88ed38012ebbfd8115d1321fa518047bc63c6626772dc"} Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.316398 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-9s82h" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.322715 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-67scw" podStartSLOduration=3.109531434 podStartE2EDuration="18.32269215s" podCreationTimestamp="2026-03-21 05:04:11 +0000 UTC" firstStartedPulling="2026-03-21 05:04:13.285801599 +0000 UTC m=+1006.262265223" lastFinishedPulling="2026-03-21 05:04:28.498962315 +0000 UTC m=+1021.475425939" observedRunningTime="2026-03-21 05:04:29.321590979 +0000 UTC m=+1022.298054603" watchObservedRunningTime="2026-03-21 05:04:29.32269215 +0000 UTC m=+1022.299155784" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.355283 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tss7r" event={"ID":"02b6af47-2c06-480b-a838-2d742efa1045","Type":"ContainerStarted","Data":"814e99ad1de41e07f65980c021d5b5f6979fbdaf273ce4c227304bbbd964feca"} Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.355788 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tss7r" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.388534 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-9g8pv" podStartSLOduration=5.326768252 podStartE2EDuration="18.388518976s" podCreationTimestamp="2026-03-21 05:04:11 +0000 UTC" firstStartedPulling="2026-03-21 05:04:12.928448781 +0000 UTC m=+1005.904912405" lastFinishedPulling="2026-03-21 05:04:25.990199485 +0000 UTC m=+1018.966663129" observedRunningTime="2026-03-21 05:04:29.385426568 +0000 UTC m=+1022.361890192" watchObservedRunningTime="2026-03-21 05:04:29.388518976 +0000 UTC m=+1022.364982590" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.433876 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-m4gqz" podStartSLOduration=2.442450467 podStartE2EDuration="18.43385771s" podCreationTimestamp="2026-03-21 05:04:11 +0000 UTC" firstStartedPulling="2026-03-21 05:04:12.483651775 +0000 UTC m=+1005.460115399" lastFinishedPulling="2026-03-21 05:04:28.475059018 +0000 UTC m=+1021.451522642" observedRunningTime="2026-03-21 05:04:29.432361627 +0000 UTC m=+1022.408825251" watchObservedRunningTime="2026-03-21 05:04:29.43385771 +0000 UTC m=+1022.410321334" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.505008 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-l59gx" podStartSLOduration=3.336834745 podStartE2EDuration="18.504993236s" podCreationTimestamp="2026-03-21 05:04:11 +0000 UTC" firstStartedPulling="2026-03-21 05:04:13.335204919 +0000 UTC m=+1006.311668543" lastFinishedPulling="2026-03-21 05:04:28.50336342 +0000 UTC m=+1021.479827034" observedRunningTime="2026-03-21 05:04:29.503639297 +0000 UTC m=+1022.480102921" watchObservedRunningTime="2026-03-21 05:04:29.504993236 +0000 UTC m=+1022.481456860" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.623316 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-7bdbc" podStartSLOduration=2.751796615 podStartE2EDuration="18.623300909s" podCreationTimestamp="2026-03-21 05:04:11 +0000 UTC" firstStartedPulling="2026-03-21 05:04:12.632650548 +0000 UTC m=+1005.609114172" lastFinishedPulling="2026-03-21 05:04:28.504154842 +0000 UTC m=+1021.480618466" observedRunningTime="2026-03-21 05:04:29.574501046 +0000 UTC m=+1022.550964670" watchObservedRunningTime="2026-03-21 05:04:29.623300909 +0000 UTC m=+1022.599764533" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.624705 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-lzvsq" podStartSLOduration=2.62357499 podStartE2EDuration="18.624698038s" podCreationTimestamp="2026-03-21 05:04:11 +0000 UTC" firstStartedPulling="2026-03-21 05:04:12.475206936 +0000 UTC m=+1005.451670560" lastFinishedPulling="2026-03-21 05:04:28.476329984 +0000 UTC m=+1021.452793608" observedRunningTime="2026-03-21 05:04:29.619444269 +0000 UTC m=+1022.595907893" watchObservedRunningTime="2026-03-21 05:04:29.624698038 +0000 UTC m=+1022.601161652" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.689608 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tss7r" podStartSLOduration=3.42278902 podStartE2EDuration="18.689589267s" podCreationTimestamp="2026-03-21 05:04:11 +0000 UTC" firstStartedPulling="2026-03-21 05:04:13.330373872 +0000 UTC m=+1006.306837506" lastFinishedPulling="2026-03-21 05:04:28.597174129 +0000 UTC m=+1021.573637753" observedRunningTime="2026-03-21 05:04:29.685470621 +0000 UTC m=+1022.661934255" watchObservedRunningTime="2026-03-21 05:04:29.689589267 +0000 UTC m=+1022.666052891" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.749715 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-9s82h" podStartSLOduration=3.519866902 podStartE2EDuration="18.749702891s" podCreationTimestamp="2026-03-21 05:04:11 +0000 UTC" firstStartedPulling="2026-03-21 05:04:13.28548837 +0000 UTC m=+1006.261951994" lastFinishedPulling="2026-03-21 05:04:28.515324359 +0000 UTC m=+1021.491787983" observedRunningTime="2026-03-21 05:04:29.749072323 +0000 UTC m=+1022.725535947" watchObservedRunningTime="2026-03-21 05:04:29.749702891 +0000 UTC m=+1022.726166505" Mar 21 05:04:29 crc kubenswrapper[4775]: I0321 05:04:29.781293 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-lxvtw" podStartSLOduration=6.000402412 podStartE2EDuration="19.781276616s" podCreationTimestamp="2026-03-21 05:04:10 +0000 UTC" firstStartedPulling="2026-03-21 05:04:12.212211153 +0000 UTC m=+1005.188674777" lastFinishedPulling="2026-03-21 05:04:25.993085347 +0000 UTC m=+1018.969548981" observedRunningTime="2026-03-21 05:04:29.77753074 +0000 UTC m=+1022.753994364" watchObservedRunningTime="2026-03-21 05:04:29.781276616 +0000 UTC m=+1022.757740240" Mar 21 05:04:32 crc kubenswrapper[4775]: I0321 05:04:32.482195 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:04:32 crc kubenswrapper[4775]: I0321 05:04:32.482750 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:04:32 crc kubenswrapper[4775]: I0321 05:04:32.482808 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 05:04:32 crc kubenswrapper[4775]: I0321 05:04:32.483365 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8d303e1558b2adccb16582693a30248fb7f96ca561d7dcb3104197e825dd15a7"} pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 05:04:32 crc kubenswrapper[4775]: I0321 05:04:32.483571 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" containerID="cri-o://8d303e1558b2adccb16582693a30248fb7f96ca561d7dcb3104197e825dd15a7" gracePeriod=600 Mar 21 05:04:33 crc kubenswrapper[4775]: I0321 05:04:33.391652 4775 generic.go:334] "Generic (PLEG): container finished" podID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerID="8d303e1558b2adccb16582693a30248fb7f96ca561d7dcb3104197e825dd15a7" exitCode=0 Mar 21 05:04:33 crc kubenswrapper[4775]: I0321 05:04:33.391705 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerDied","Data":"8d303e1558b2adccb16582693a30248fb7f96ca561d7dcb3104197e825dd15a7"} Mar 21 05:04:33 crc kubenswrapper[4775]: I0321 05:04:33.391747 4775 scope.go:117] "RemoveContainer" containerID="eaf646a6237d4b4ce6a8a82755505f01a336cf93fef407f02dbf82d68f5008b4" Mar 21 05:04:35 crc kubenswrapper[4775]: I0321 05:04:35.409415 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jvglf" event={"ID":"95a44b12-e027-400d-b257-99f2012251d8","Type":"ContainerStarted","Data":"b9109d8ccbf32cbfa0cff8c61c0680d0eb7eb54e6d830c96fe75217312b7cdc8"} Mar 21 05:04:35 crc kubenswrapper[4775]: I0321 05:04:35.411793 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-767865f676-4tbvg" event={"ID":"9bfe7d25-53ea-484d-a481-0ea04ee2b8a8","Type":"ContainerStarted","Data":"8d16199ef493b24538615a53998624f122687acb65de5bf87a5779df74d2e4de"} Mar 21 05:04:35 crc kubenswrapper[4775]: I0321 05:04:35.412356 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-767865f676-4tbvg" Mar 21 05:04:35 crc kubenswrapper[4775]: I0321 05:04:35.414290 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-gdrc5" event={"ID":"3c9f18bd-def6-45ff-a92b-25c6f40d6bb5","Type":"ContainerStarted","Data":"451c6ed00601aa4e554b35fe9e8592269a0eeafa83748f0324e01ae4f1e7fdd3"} Mar 21 05:04:35 crc kubenswrapper[4775]: I0321 05:04:35.415357 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-gdrc5" Mar 21 05:04:35 crc kubenswrapper[4775]: I0321 05:04:35.423683 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerStarted","Data":"09c71a6e96dc622a58adf5b83a67eab26ff45301d2bfcc2a43f1cbd9eb2d9791"} Mar 21 05:04:35 crc kubenswrapper[4775]: I0321 05:04:35.426452 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hf688" event={"ID":"ec38d53e-6fe4-41b5-8548-e49fadd9d6bf","Type":"ContainerStarted","Data":"6ac8d07514471eaf2df48765e096536120a65a71579498a3b896067f295f9f69"} Mar 21 05:04:35 crc kubenswrapper[4775]: I0321 05:04:35.426734 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hf688" Mar 21 05:04:35 crc kubenswrapper[4775]: I0321 05:04:35.431785 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-884679f54-ctm9h" event={"ID":"6eeb04ad-7251-488c-bd52-b2f14f6fb68b","Type":"ContainerStarted","Data":"9705459b49ed17e605e6a92451d276563775b14e8862b733af598edd781b6505"} Mar 21 05:04:35 crc kubenswrapper[4775]: I0321 05:04:35.432020 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-884679f54-ctm9h" Mar 21 05:04:35 crc kubenswrapper[4775]: I0321 05:04:35.439145 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-55f864c847-jq88h" podStartSLOduration=8.906206983 podStartE2EDuration="24.439109781s" podCreationTimestamp="2026-03-21 05:04:11 +0000 UTC" firstStartedPulling="2026-03-21 05:04:12.943598741 +0000 UTC m=+1005.920062365" lastFinishedPulling="2026-03-21 05:04:28.476501539 +0000 UTC m=+1021.452965163" observedRunningTime="2026-03-21 05:04:29.811391369 +0000 UTC m=+1022.787854993" watchObservedRunningTime="2026-03-21 05:04:35.439109781 +0000 UTC m=+1028.415573415" Mar 21 05:04:35 crc kubenswrapper[4775]: I0321 05:04:35.452461 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-884679f54-ctm9h" podStartSLOduration=3.138134134 podStartE2EDuration="24.452443349s" podCreationTimestamp="2026-03-21 05:04:11 +0000 UTC" firstStartedPulling="2026-03-21 05:04:13.351536662 +0000 UTC m=+1006.328000286" lastFinishedPulling="2026-03-21 05:04:34.665845877 +0000 UTC m=+1027.642309501" observedRunningTime="2026-03-21 05:04:35.448871448 +0000 UTC m=+1028.425335092" watchObservedRunningTime="2026-03-21 05:04:35.452443349 +0000 UTC m=+1028.428906993" Mar 21 05:04:35 crc kubenswrapper[4775]: I0321 05:04:35.453282 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jvglf" podStartSLOduration=3.054208466 podStartE2EDuration="24.453276763s" podCreationTimestamp="2026-03-21 05:04:11 +0000 UTC" firstStartedPulling="2026-03-21 05:04:13.351381317 +0000 UTC m=+1006.327844941" lastFinishedPulling="2026-03-21 05:04:34.750449594 +0000 UTC m=+1027.726913238" observedRunningTime="2026-03-21 05:04:35.43764128 +0000 UTC m=+1028.414104904" watchObservedRunningTime="2026-03-21 05:04:35.453276763 +0000 UTC m=+1028.429740397" Mar 21 05:04:35 crc kubenswrapper[4775]: I0321 05:04:35.490295 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-gdrc5" podStartSLOduration=3.193918865 podStartE2EDuration="24.490276851s" podCreationTimestamp="2026-03-21 05:04:11 +0000 UTC" firstStartedPulling="2026-03-21 05:04:13.368200604 +0000 UTC m=+1006.344664228" lastFinishedPulling="2026-03-21 05:04:34.6645586 +0000 UTC m=+1027.641022214" observedRunningTime="2026-03-21 05:04:35.471803458 +0000 UTC m=+1028.448267102" watchObservedRunningTime="2026-03-21 05:04:35.490276851 +0000 UTC m=+1028.466740475" Mar 21 05:04:35 crc kubenswrapper[4775]: I0321 05:04:35.490867 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-767865f676-4tbvg" podStartSLOduration=3.185789635 podStartE2EDuration="24.490861738s" podCreationTimestamp="2026-03-21 05:04:11 +0000 UTC" firstStartedPulling="2026-03-21 05:04:13.35149286 +0000 UTC m=+1006.327956484" lastFinishedPulling="2026-03-21 05:04:34.656564963 +0000 UTC m=+1027.633028587" observedRunningTime="2026-03-21 05:04:35.488645905 +0000 UTC m=+1028.465109539" watchObservedRunningTime="2026-03-21 05:04:35.490861738 +0000 UTC m=+1028.467325362" Mar 21 05:04:35 crc kubenswrapper[4775]: I0321 05:04:35.531054 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hf688" podStartSLOduration=3.226752787 podStartE2EDuration="24.531032797s" podCreationTimestamp="2026-03-21 05:04:11 +0000 UTC" firstStartedPulling="2026-03-21 05:04:13.351640975 +0000 UTC m=+1006.328104599" lastFinishedPulling="2026-03-21 05:04:34.655920985 +0000 UTC m=+1027.632384609" observedRunningTime="2026-03-21 05:04:35.525637474 +0000 UTC m=+1028.502101118" watchObservedRunningTime="2026-03-21 05:04:35.531032797 +0000 UTC m=+1028.507496421" Mar 21 05:04:37 crc kubenswrapper[4775]: I0321 05:04:37.405760 4775 scope.go:117] "RemoveContainer" containerID="a6f778f857c6aa3ee7f297014df3be268b9de30de7297b84506cc38e86b9d649" Mar 21 05:04:41 crc kubenswrapper[4775]: I0321 05:04:41.342435 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-twhxx" Mar 21 05:04:41 crc kubenswrapper[4775]: I0321 05:04:41.343197 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-dn22m" Mar 21 05:04:41 crc kubenswrapper[4775]: I0321 05:04:41.384168 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-lxvtw" Mar 21 05:04:41 crc kubenswrapper[4775]: I0321 05:04:41.404900 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-lzvsq" Mar 21 05:04:41 crc kubenswrapper[4775]: I0321 05:04:41.445377 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-7bdbc" Mar 21 05:04:41 crc kubenswrapper[4775]: I0321 05:04:41.446403 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-m4gqz" Mar 21 05:04:41 crc kubenswrapper[4775]: I0321 05:04:41.549262 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-9g8pv" Mar 21 05:04:41 crc kubenswrapper[4775]: I0321 05:04:41.719001 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-55f864c847-jq88h" Mar 21 05:04:41 crc kubenswrapper[4775]: I0321 05:04:41.749364 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-hf688" Mar 21 05:04:41 crc kubenswrapper[4775]: I0321 05:04:41.759807 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-767865f676-4tbvg" Mar 21 05:04:41 crc kubenswrapper[4775]: I0321 05:04:41.801837 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tss7r" Mar 21 05:04:41 crc kubenswrapper[4775]: I0321 05:04:41.826487 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-67scw" Mar 21 05:04:41 crc kubenswrapper[4775]: I0321 05:04:41.895954 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5784578c99-jj4pt" Mar 21 05:04:41 crc kubenswrapper[4775]: I0321 05:04:41.941282 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-884679f54-ctm9h" Mar 21 05:04:41 crc kubenswrapper[4775]: I0321 05:04:41.973973 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-c674c5965-lqmgv" Mar 21 05:04:42 crc kubenswrapper[4775]: I0321 05:04:42.084903 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-9s82h" Mar 21 05:04:42 crc kubenswrapper[4775]: I0321 05:04:42.100665 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-l59gx" Mar 21 05:04:42 crc kubenswrapper[4775]: I0321 05:04:42.208502 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-gdrc5" Mar 21 05:04:43 crc kubenswrapper[4775]: I0321 05:04:43.167940 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert\") pod \"infra-operator-controller-manager-65f65cc49c-2mgp8\" (UID: \"0a66456f-7860-4dc1-9c1c-0db69ddcc800\") " pod="openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8" Mar 21 05:04:43 crc kubenswrapper[4775]: I0321 05:04:43.182241 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0a66456f-7860-4dc1-9c1c-0db69ddcc800-cert\") pod \"infra-operator-controller-manager-65f65cc49c-2mgp8\" (UID: \"0a66456f-7860-4dc1-9c1c-0db69ddcc800\") " pod="openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8" Mar 21 05:04:43 crc kubenswrapper[4775]: I0321 05:04:43.275738 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-fdr25" Mar 21 05:04:43 crc kubenswrapper[4775]: I0321 05:04:43.284877 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8" Mar 21 05:04:43 crc kubenswrapper[4775]: I0321 05:04:43.476663 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rwq59\" (UID: \"3035739a-202f-4794-bb4f-ae2342a96441\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59" Mar 21 05:04:43 crc kubenswrapper[4775]: I0321 05:04:43.482537 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3035739a-202f-4794-bb4f-ae2342a96441-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rwq59\" (UID: \"3035739a-202f-4794-bb4f-ae2342a96441\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59" Mar 21 05:04:43 crc kubenswrapper[4775]: I0321 05:04:43.658141 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-cthp5" Mar 21 05:04:43 crc kubenswrapper[4775]: I0321 05:04:43.665421 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59" Mar 21 05:04:43 crc kubenswrapper[4775]: W0321 05:04:43.700177 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a66456f_7860_4dc1_9c1c_0db69ddcc800.slice/crio-f3476740abd8ffe4ea69aa39b1d69aa40ae063bd15f785b821410788c847f2e1 WatchSource:0}: Error finding container f3476740abd8ffe4ea69aa39b1d69aa40ae063bd15f785b821410788c847f2e1: Status 404 returned error can't find the container with id f3476740abd8ffe4ea69aa39b1d69aa40ae063bd15f785b821410788c847f2e1 Mar 21 05:04:43 crc kubenswrapper[4775]: I0321 05:04:43.701586 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8"] Mar 21 05:04:43 crc kubenswrapper[4775]: I0321 05:04:43.781515 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:43 crc kubenswrapper[4775]: I0321 05:04:43.781700 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:43 crc kubenswrapper[4775]: I0321 05:04:43.790924 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-metrics-certs\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:43 crc kubenswrapper[4775]: I0321 05:04:43.793860 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/907f0cdf-2d87-4d09-97af-5591d061b4f6-webhook-certs\") pod \"openstack-operator-controller-manager-65746ff4dc-hg4rq\" (UID: \"907f0cdf-2d87-4d09-97af-5591d061b4f6\") " pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:43 crc kubenswrapper[4775]: I0321 05:04:43.891480 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-zqnnz" Mar 21 05:04:43 crc kubenswrapper[4775]: I0321 05:04:43.899896 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:44 crc kubenswrapper[4775]: I0321 05:04:44.095839 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59"] Mar 21 05:04:44 crc kubenswrapper[4775]: I0321 05:04:44.126326 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq"] Mar 21 05:04:44 crc kubenswrapper[4775]: W0321 05:04:44.126381 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod907f0cdf_2d87_4d09_97af_5591d061b4f6.slice/crio-0b21472eea31b0e2d7f39cfe5e58d32c32feb754ddcc17473770ec2ea521949c WatchSource:0}: Error finding container 0b21472eea31b0e2d7f39cfe5e58d32c32feb754ddcc17473770ec2ea521949c: Status 404 returned error can't find the container with id 0b21472eea31b0e2d7f39cfe5e58d32c32feb754ddcc17473770ec2ea521949c Mar 21 05:04:44 crc kubenswrapper[4775]: I0321 05:04:44.505497 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" event={"ID":"907f0cdf-2d87-4d09-97af-5591d061b4f6","Type":"ContainerStarted","Data":"0b21472eea31b0e2d7f39cfe5e58d32c32feb754ddcc17473770ec2ea521949c"} Mar 21 05:04:44 crc kubenswrapper[4775]: I0321 05:04:44.507568 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8" event={"ID":"0a66456f-7860-4dc1-9c1c-0db69ddcc800","Type":"ContainerStarted","Data":"f3476740abd8ffe4ea69aa39b1d69aa40ae063bd15f785b821410788c847f2e1"} Mar 21 05:04:44 crc kubenswrapper[4775]: I0321 05:04:44.508962 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59" event={"ID":"3035739a-202f-4794-bb4f-ae2342a96441","Type":"ContainerStarted","Data":"aaf4afecb3a24559f5e03ae7f8304d8221883557cd03befa4b7f56c9f9157e02"} Mar 21 05:04:49 crc kubenswrapper[4775]: I0321 05:04:49.547106 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" event={"ID":"907f0cdf-2d87-4d09-97af-5591d061b4f6","Type":"ContainerStarted","Data":"1a4bea49d6fe2de12f4e42fa90462b35cc170947f406707e3b3cac159711d6cd"} Mar 21 05:04:50 crc kubenswrapper[4775]: I0321 05:04:50.552333 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:04:50 crc kubenswrapper[4775]: I0321 05:04:50.585315 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" podStartSLOduration=39.585295485 podStartE2EDuration="39.585295485s" podCreationTimestamp="2026-03-21 05:04:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:04:50.582765003 +0000 UTC m=+1043.559228667" watchObservedRunningTime="2026-03-21 05:04:50.585295485 +0000 UTC m=+1043.561759109" Mar 21 05:04:53 crc kubenswrapper[4775]: I0321 05:04:53.573808 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8" event={"ID":"0a66456f-7860-4dc1-9c1c-0db69ddcc800","Type":"ContainerStarted","Data":"2288ecffa8b7fc13461f7bb72cc6b3b81443765bc5a9cb5baf4b63c40a61978f"} Mar 21 05:04:53 crc kubenswrapper[4775]: I0321 05:04:53.575421 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8" Mar 21 05:04:53 crc kubenswrapper[4775]: I0321 05:04:53.577340 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59" event={"ID":"3035739a-202f-4794-bb4f-ae2342a96441","Type":"ContainerStarted","Data":"6ddf9d8507a99c5f6069ed20292ddc2ddd0d77f55a16a4db939eb0792bc68880"} Mar 21 05:04:53 crc kubenswrapper[4775]: I0321 05:04:53.578055 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59" Mar 21 05:04:53 crc kubenswrapper[4775]: I0321 05:04:53.580869 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c7pjw" event={"ID":"49a3c6c7-9e86-495d-8d1a-486d6bfbbbdd","Type":"ContainerStarted","Data":"8fad59adcb09e434567ccb04bde888708623627b90fef5c5cb347ba318a1911d"} Mar 21 05:04:53 crc kubenswrapper[4775]: I0321 05:04:53.581205 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c7pjw" Mar 21 05:04:53 crc kubenswrapper[4775]: I0321 05:04:53.617665 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c7pjw" podStartSLOduration=2.89936375 podStartE2EDuration="42.617649031s" podCreationTimestamp="2026-03-21 05:04:11 +0000 UTC" firstStartedPulling="2026-03-21 05:04:12.92805004 +0000 UTC m=+1005.904513664" lastFinishedPulling="2026-03-21 05:04:52.646335321 +0000 UTC m=+1045.622798945" observedRunningTime="2026-03-21 05:04:53.615331945 +0000 UTC m=+1046.591795569" watchObservedRunningTime="2026-03-21 05:04:53.617649031 +0000 UTC m=+1046.594112655" Mar 21 05:04:53 crc kubenswrapper[4775]: I0321 05:04:53.620024 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8" podStartSLOduration=33.642300307 podStartE2EDuration="42.620016948s" podCreationTimestamp="2026-03-21 05:04:11 +0000 UTC" firstStartedPulling="2026-03-21 05:04:43.704537096 +0000 UTC m=+1036.681000760" lastFinishedPulling="2026-03-21 05:04:52.682253777 +0000 UTC m=+1045.658717401" observedRunningTime="2026-03-21 05:04:53.591835891 +0000 UTC m=+1046.568299525" watchObservedRunningTime="2026-03-21 05:04:53.620016948 +0000 UTC m=+1046.596480582" Mar 21 05:04:53 crc kubenswrapper[4775]: I0321 05:04:53.641998 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59" podStartSLOduration=34.068200405 podStartE2EDuration="42.641978849s" podCreationTimestamp="2026-03-21 05:04:11 +0000 UTC" firstStartedPulling="2026-03-21 05:04:44.10766935 +0000 UTC m=+1037.084132974" lastFinishedPulling="2026-03-21 05:04:52.681447794 +0000 UTC m=+1045.657911418" observedRunningTime="2026-03-21 05:04:53.637761299 +0000 UTC m=+1046.614224923" watchObservedRunningTime="2026-03-21 05:04:53.641978849 +0000 UTC m=+1046.618442473" Mar 21 05:05:01 crc kubenswrapper[4775]: I0321 05:05:01.579653 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c7pjw" Mar 21 05:05:03 crc kubenswrapper[4775]: I0321 05:05:03.293760 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-65f65cc49c-2mgp8" Mar 21 05:05:03 crc kubenswrapper[4775]: I0321 05:05:03.679187 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rwq59" Mar 21 05:05:03 crc kubenswrapper[4775]: I0321 05:05:03.912437 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-65746ff4dc-hg4rq" Mar 21 05:05:19 crc kubenswrapper[4775]: I0321 05:05:19.774604 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-d942p"] Mar 21 05:05:19 crc kubenswrapper[4775]: E0321 05:05:19.776376 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="694f1c66-ce56-4854-9f4c-02a42e16046d" containerName="extract-utilities" Mar 21 05:05:19 crc kubenswrapper[4775]: I0321 05:05:19.776398 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="694f1c66-ce56-4854-9f4c-02a42e16046d" containerName="extract-utilities" Mar 21 05:05:19 crc kubenswrapper[4775]: E0321 05:05:19.776414 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="694f1c66-ce56-4854-9f4c-02a42e16046d" containerName="registry-server" Mar 21 05:05:19 crc kubenswrapper[4775]: I0321 05:05:19.776422 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="694f1c66-ce56-4854-9f4c-02a42e16046d" containerName="registry-server" Mar 21 05:05:19 crc kubenswrapper[4775]: E0321 05:05:19.776440 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="694f1c66-ce56-4854-9f4c-02a42e16046d" containerName="extract-content" Mar 21 05:05:19 crc kubenswrapper[4775]: I0321 05:05:19.776448 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="694f1c66-ce56-4854-9f4c-02a42e16046d" containerName="extract-content" Mar 21 05:05:19 crc kubenswrapper[4775]: I0321 05:05:19.776657 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="694f1c66-ce56-4854-9f4c-02a42e16046d" containerName="registry-server" Mar 21 05:05:19 crc kubenswrapper[4775]: I0321 05:05:19.777564 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-d942p" Mar 21 05:05:19 crc kubenswrapper[4775]: I0321 05:05:19.779411 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Mar 21 05:05:19 crc kubenswrapper[4775]: I0321 05:05:19.779623 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-sndkl" Mar 21 05:05:19 crc kubenswrapper[4775]: I0321 05:05:19.780548 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Mar 21 05:05:19 crc kubenswrapper[4775]: I0321 05:05:19.781137 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Mar 21 05:05:19 crc kubenswrapper[4775]: I0321 05:05:19.792376 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-d942p"] Mar 21 05:05:19 crc kubenswrapper[4775]: I0321 05:05:19.827330 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bqfch"] Mar 21 05:05:19 crc kubenswrapper[4775]: I0321 05:05:19.828610 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-bqfch" Mar 21 05:05:19 crc kubenswrapper[4775]: I0321 05:05:19.835471 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Mar 21 05:05:19 crc kubenswrapper[4775]: I0321 05:05:19.843476 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bqfch"] Mar 21 05:05:19 crc kubenswrapper[4775]: I0321 05:05:19.968973 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8c11166-8b08-43d6-84e3-e5cbd0adbd8d-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-bqfch\" (UID: \"e8c11166-8b08-43d6-84e3-e5cbd0adbd8d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bqfch" Mar 21 05:05:19 crc kubenswrapper[4775]: I0321 05:05:19.969328 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62c4e9af-66f9-4f32-8c9a-30d4907a753c-config\") pod \"dnsmasq-dns-675f4bcbfc-d942p\" (UID: \"62c4e9af-66f9-4f32-8c9a-30d4907a753c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-d942p" Mar 21 05:05:19 crc kubenswrapper[4775]: I0321 05:05:19.969431 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74n7g\" (UniqueName: \"kubernetes.io/projected/e8c11166-8b08-43d6-84e3-e5cbd0adbd8d-kube-api-access-74n7g\") pod \"dnsmasq-dns-78dd6ddcc-bqfch\" (UID: \"e8c11166-8b08-43d6-84e3-e5cbd0adbd8d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bqfch" Mar 21 05:05:19 crc kubenswrapper[4775]: I0321 05:05:19.969805 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8c11166-8b08-43d6-84e3-e5cbd0adbd8d-config\") pod \"dnsmasq-dns-78dd6ddcc-bqfch\" (UID: \"e8c11166-8b08-43d6-84e3-e5cbd0adbd8d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bqfch" Mar 21 05:05:19 crc kubenswrapper[4775]: I0321 05:05:19.970028 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlpch\" (UniqueName: \"kubernetes.io/projected/62c4e9af-66f9-4f32-8c9a-30d4907a753c-kube-api-access-vlpch\") pod \"dnsmasq-dns-675f4bcbfc-d942p\" (UID: \"62c4e9af-66f9-4f32-8c9a-30d4907a753c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-d942p" Mar 21 05:05:20 crc kubenswrapper[4775]: I0321 05:05:20.072084 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8c11166-8b08-43d6-84e3-e5cbd0adbd8d-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-bqfch\" (UID: \"e8c11166-8b08-43d6-84e3-e5cbd0adbd8d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bqfch" Mar 21 05:05:20 crc kubenswrapper[4775]: I0321 05:05:20.072464 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62c4e9af-66f9-4f32-8c9a-30d4907a753c-config\") pod \"dnsmasq-dns-675f4bcbfc-d942p\" (UID: \"62c4e9af-66f9-4f32-8c9a-30d4907a753c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-d942p" Mar 21 05:05:20 crc kubenswrapper[4775]: I0321 05:05:20.072487 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74n7g\" (UniqueName: \"kubernetes.io/projected/e8c11166-8b08-43d6-84e3-e5cbd0adbd8d-kube-api-access-74n7g\") pod \"dnsmasq-dns-78dd6ddcc-bqfch\" (UID: \"e8c11166-8b08-43d6-84e3-e5cbd0adbd8d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bqfch" Mar 21 05:05:20 crc kubenswrapper[4775]: I0321 05:05:20.072852 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8c11166-8b08-43d6-84e3-e5cbd0adbd8d-config\") pod \"dnsmasq-dns-78dd6ddcc-bqfch\" (UID: \"e8c11166-8b08-43d6-84e3-e5cbd0adbd8d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bqfch" Mar 21 05:05:20 crc kubenswrapper[4775]: I0321 05:05:20.073080 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlpch\" (UniqueName: \"kubernetes.io/projected/62c4e9af-66f9-4f32-8c9a-30d4907a753c-kube-api-access-vlpch\") pod \"dnsmasq-dns-675f4bcbfc-d942p\" (UID: \"62c4e9af-66f9-4f32-8c9a-30d4907a753c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-d942p" Mar 21 05:05:20 crc kubenswrapper[4775]: I0321 05:05:20.073423 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62c4e9af-66f9-4f32-8c9a-30d4907a753c-config\") pod \"dnsmasq-dns-675f4bcbfc-d942p\" (UID: \"62c4e9af-66f9-4f32-8c9a-30d4907a753c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-d942p" Mar 21 05:05:20 crc kubenswrapper[4775]: I0321 05:05:20.073740 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8c11166-8b08-43d6-84e3-e5cbd0adbd8d-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-bqfch\" (UID: \"e8c11166-8b08-43d6-84e3-e5cbd0adbd8d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bqfch" Mar 21 05:05:20 crc kubenswrapper[4775]: I0321 05:05:20.074427 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8c11166-8b08-43d6-84e3-e5cbd0adbd8d-config\") pod \"dnsmasq-dns-78dd6ddcc-bqfch\" (UID: \"e8c11166-8b08-43d6-84e3-e5cbd0adbd8d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bqfch" Mar 21 05:05:20 crc kubenswrapper[4775]: I0321 05:05:20.095203 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlpch\" (UniqueName: \"kubernetes.io/projected/62c4e9af-66f9-4f32-8c9a-30d4907a753c-kube-api-access-vlpch\") pod \"dnsmasq-dns-675f4bcbfc-d942p\" (UID: \"62c4e9af-66f9-4f32-8c9a-30d4907a753c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-d942p" Mar 21 05:05:20 crc kubenswrapper[4775]: I0321 05:05:20.101886 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74n7g\" (UniqueName: \"kubernetes.io/projected/e8c11166-8b08-43d6-84e3-e5cbd0adbd8d-kube-api-access-74n7g\") pod \"dnsmasq-dns-78dd6ddcc-bqfch\" (UID: \"e8c11166-8b08-43d6-84e3-e5cbd0adbd8d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bqfch" Mar 21 05:05:20 crc kubenswrapper[4775]: I0321 05:05:20.103983 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-d942p" Mar 21 05:05:20 crc kubenswrapper[4775]: I0321 05:05:20.156108 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-bqfch" Mar 21 05:05:20 crc kubenswrapper[4775]: I0321 05:05:20.595456 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-d942p"] Mar 21 05:05:20 crc kubenswrapper[4775]: I0321 05:05:20.634957 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bqfch"] Mar 21 05:05:20 crc kubenswrapper[4775]: I0321 05:05:20.762474 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-bqfch" event={"ID":"e8c11166-8b08-43d6-84e3-e5cbd0adbd8d","Type":"ContainerStarted","Data":"59274326bf041a9ff442e3ad650874c430fb6d402b02a40d2d8ac6b554c247ca"} Mar 21 05:05:20 crc kubenswrapper[4775]: I0321 05:05:20.763825 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-d942p" event={"ID":"62c4e9af-66f9-4f32-8c9a-30d4907a753c","Type":"ContainerStarted","Data":"a875cb7a582da7fe33690731a6ec7fd7d012b80f676cca55452c73ba7b74cfd0"} Mar 21 05:05:22 crc kubenswrapper[4775]: I0321 05:05:22.722765 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-d942p"] Mar 21 05:05:22 crc kubenswrapper[4775]: I0321 05:05:22.741365 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-4rt6c"] Mar 21 05:05:22 crc kubenswrapper[4775]: I0321 05:05:22.742461 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" Mar 21 05:05:22 crc kubenswrapper[4775]: I0321 05:05:22.763283 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-4rt6c"] Mar 21 05:05:22 crc kubenswrapper[4775]: I0321 05:05:22.817808 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b9920a8-df59-4747-8f72-1f25d22517b3-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-4rt6c\" (UID: \"6b9920a8-df59-4747-8f72-1f25d22517b3\") " pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" Mar 21 05:05:22 crc kubenswrapper[4775]: I0321 05:05:22.817890 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b9920a8-df59-4747-8f72-1f25d22517b3-config\") pod \"dnsmasq-dns-5ccc8479f9-4rt6c\" (UID: \"6b9920a8-df59-4747-8f72-1f25d22517b3\") " pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" Mar 21 05:05:22 crc kubenswrapper[4775]: I0321 05:05:22.817925 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zth55\" (UniqueName: \"kubernetes.io/projected/6b9920a8-df59-4747-8f72-1f25d22517b3-kube-api-access-zth55\") pod \"dnsmasq-dns-5ccc8479f9-4rt6c\" (UID: \"6b9920a8-df59-4747-8f72-1f25d22517b3\") " pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" Mar 21 05:05:22 crc kubenswrapper[4775]: I0321 05:05:22.918751 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zth55\" (UniqueName: \"kubernetes.io/projected/6b9920a8-df59-4747-8f72-1f25d22517b3-kube-api-access-zth55\") pod \"dnsmasq-dns-5ccc8479f9-4rt6c\" (UID: \"6b9920a8-df59-4747-8f72-1f25d22517b3\") " pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" Mar 21 05:05:22 crc kubenswrapper[4775]: I0321 05:05:22.918852 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b9920a8-df59-4747-8f72-1f25d22517b3-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-4rt6c\" (UID: \"6b9920a8-df59-4747-8f72-1f25d22517b3\") " pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" Mar 21 05:05:22 crc kubenswrapper[4775]: I0321 05:05:22.918879 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b9920a8-df59-4747-8f72-1f25d22517b3-config\") pod \"dnsmasq-dns-5ccc8479f9-4rt6c\" (UID: \"6b9920a8-df59-4747-8f72-1f25d22517b3\") " pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" Mar 21 05:05:22 crc kubenswrapper[4775]: I0321 05:05:22.919732 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b9920a8-df59-4747-8f72-1f25d22517b3-config\") pod \"dnsmasq-dns-5ccc8479f9-4rt6c\" (UID: \"6b9920a8-df59-4747-8f72-1f25d22517b3\") " pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" Mar 21 05:05:22 crc kubenswrapper[4775]: I0321 05:05:22.919748 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b9920a8-df59-4747-8f72-1f25d22517b3-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-4rt6c\" (UID: \"6b9920a8-df59-4747-8f72-1f25d22517b3\") " pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" Mar 21 05:05:22 crc kubenswrapper[4775]: I0321 05:05:22.945256 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zth55\" (UniqueName: \"kubernetes.io/projected/6b9920a8-df59-4747-8f72-1f25d22517b3-kube-api-access-zth55\") pod \"dnsmasq-dns-5ccc8479f9-4rt6c\" (UID: \"6b9920a8-df59-4747-8f72-1f25d22517b3\") " pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.004191 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bqfch"] Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.026205 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mn5xb"] Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.027332 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.032966 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mn5xb"] Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.085383 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.226498 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f77fw\" (UniqueName: \"kubernetes.io/projected/2084b413-8c86-4cc0-89ce-0dbfc4049e9b-kube-api-access-f77fw\") pod \"dnsmasq-dns-57d769cc4f-mn5xb\" (UID: \"2084b413-8c86-4cc0-89ce-0dbfc4049e9b\") " pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.226828 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2084b413-8c86-4cc0-89ce-0dbfc4049e9b-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-mn5xb\" (UID: \"2084b413-8c86-4cc0-89ce-0dbfc4049e9b\") " pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.226887 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2084b413-8c86-4cc0-89ce-0dbfc4049e9b-config\") pod \"dnsmasq-dns-57d769cc4f-mn5xb\" (UID: \"2084b413-8c86-4cc0-89ce-0dbfc4049e9b\") " pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.328437 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f77fw\" (UniqueName: \"kubernetes.io/projected/2084b413-8c86-4cc0-89ce-0dbfc4049e9b-kube-api-access-f77fw\") pod \"dnsmasq-dns-57d769cc4f-mn5xb\" (UID: \"2084b413-8c86-4cc0-89ce-0dbfc4049e9b\") " pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.328570 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2084b413-8c86-4cc0-89ce-0dbfc4049e9b-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-mn5xb\" (UID: \"2084b413-8c86-4cc0-89ce-0dbfc4049e9b\") " pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.328611 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2084b413-8c86-4cc0-89ce-0dbfc4049e9b-config\") pod \"dnsmasq-dns-57d769cc4f-mn5xb\" (UID: \"2084b413-8c86-4cc0-89ce-0dbfc4049e9b\") " pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.330473 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2084b413-8c86-4cc0-89ce-0dbfc4049e9b-config\") pod \"dnsmasq-dns-57d769cc4f-mn5xb\" (UID: \"2084b413-8c86-4cc0-89ce-0dbfc4049e9b\") " pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.330634 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2084b413-8c86-4cc0-89ce-0dbfc4049e9b-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-mn5xb\" (UID: \"2084b413-8c86-4cc0-89ce-0dbfc4049e9b\") " pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.367988 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f77fw\" (UniqueName: \"kubernetes.io/projected/2084b413-8c86-4cc0-89ce-0dbfc4049e9b-kube-api-access-f77fw\") pod \"dnsmasq-dns-57d769cc4f-mn5xb\" (UID: \"2084b413-8c86-4cc0-89ce-0dbfc4049e9b\") " pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.609037 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-4rt6c"] Mar 21 05:05:23 crc kubenswrapper[4775]: W0321 05:05:23.621770 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b9920a8_df59_4747_8f72_1f25d22517b3.slice/crio-70d41f601713d43b2beb92b797cb5b68ffe82649d11afaeaa30ae3ca5021d995 WatchSource:0}: Error finding container 70d41f601713d43b2beb92b797cb5b68ffe82649d11afaeaa30ae3ca5021d995: Status 404 returned error can't find the container with id 70d41f601713d43b2beb92b797cb5b68ffe82649d11afaeaa30ae3ca5021d995 Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.642838 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.808180 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" event={"ID":"6b9920a8-df59-4747-8f72-1f25d22517b3","Type":"ContainerStarted","Data":"70d41f601713d43b2beb92b797cb5b68ffe82649d11afaeaa30ae3ca5021d995"} Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.874052 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.877568 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.881648 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.881807 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.881838 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.881968 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.881980 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-5pn7h" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.882263 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.886214 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 21 05:05:23 crc kubenswrapper[4775]: I0321 05:05:23.897551 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.037751 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/375fb8b7-b673-4fd7-ae51-5f82f33c196f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.038108 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.038174 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.038203 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/375fb8b7-b673-4fd7-ae51-5f82f33c196f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.038230 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/375fb8b7-b673-4fd7-ae51-5f82f33c196f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.038273 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.038298 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.038349 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqdt6\" (UniqueName: \"kubernetes.io/projected/375fb8b7-b673-4fd7-ae51-5f82f33c196f-kube-api-access-vqdt6\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.038395 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/375fb8b7-b673-4fd7-ae51-5f82f33c196f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.038490 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.038557 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/375fb8b7-b673-4fd7-ae51-5f82f33c196f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.139892 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/375fb8b7-b673-4fd7-ae51-5f82f33c196f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.139944 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/375fb8b7-b673-4fd7-ae51-5f82f33c196f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.139987 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.140008 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.140022 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/375fb8b7-b673-4fd7-ae51-5f82f33c196f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.140048 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/375fb8b7-b673-4fd7-ae51-5f82f33c196f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.140068 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.140083 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.140103 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqdt6\" (UniqueName: \"kubernetes.io/projected/375fb8b7-b673-4fd7-ae51-5f82f33c196f-kube-api-access-vqdt6\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.140152 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/375fb8b7-b673-4fd7-ae51-5f82f33c196f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.140175 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.140541 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.141268 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.141403 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/375fb8b7-b673-4fd7-ae51-5f82f33c196f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.141696 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/375fb8b7-b673-4fd7-ae51-5f82f33c196f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.143319 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/375fb8b7-b673-4fd7-ae51-5f82f33c196f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.143948 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.153687 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/375fb8b7-b673-4fd7-ae51-5f82f33c196f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.153753 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.154195 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.163772 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/375fb8b7-b673-4fd7-ae51-5f82f33c196f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.168234 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mn5xb"] Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.170356 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqdt6\" (UniqueName: \"kubernetes.io/projected/375fb8b7-b673-4fd7-ae51-5f82f33c196f-kube-api-access-vqdt6\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.175081 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.178883 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.181391 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.181813 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.181961 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-kmjnx" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.182253 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.182610 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.184528 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.184755 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.194485 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.207341 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.343483 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/839e915e-8197-48e9-8b69-56ac420a1eed-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.343542 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.343576 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmtzj\" (UniqueName: \"kubernetes.io/projected/839e915e-8197-48e9-8b69-56ac420a1eed-kube-api-access-lmtzj\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.343613 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.343682 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/839e915e-8197-48e9-8b69-56ac420a1eed-server-conf\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.343731 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.343770 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/839e915e-8197-48e9-8b69-56ac420a1eed-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.343856 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/839e915e-8197-48e9-8b69-56ac420a1eed-config-data\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.343904 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.343933 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.343996 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/839e915e-8197-48e9-8b69-56ac420a1eed-pod-info\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.445552 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.445596 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmtzj\" (UniqueName: \"kubernetes.io/projected/839e915e-8197-48e9-8b69-56ac420a1eed-kube-api-access-lmtzj\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.445628 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.445652 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/839e915e-8197-48e9-8b69-56ac420a1eed-server-conf\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.445677 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.445717 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/839e915e-8197-48e9-8b69-56ac420a1eed-config-data\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.445733 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/839e915e-8197-48e9-8b69-56ac420a1eed-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.445752 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.445768 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.445808 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/839e915e-8197-48e9-8b69-56ac420a1eed-pod-info\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.445841 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/839e915e-8197-48e9-8b69-56ac420a1eed-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.446168 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.446676 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.446805 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.446874 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/839e915e-8197-48e9-8b69-56ac420a1eed-config-data\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.447411 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/839e915e-8197-48e9-8b69-56ac420a1eed-server-conf\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.448023 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/839e915e-8197-48e9-8b69-56ac420a1eed-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.451311 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.452704 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.457882 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/839e915e-8197-48e9-8b69-56ac420a1eed-pod-info\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.464012 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/839e915e-8197-48e9-8b69-56ac420a1eed-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.464467 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmtzj\" (UniqueName: \"kubernetes.io/projected/839e915e-8197-48e9-8b69-56ac420a1eed-kube-api-access-lmtzj\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.479095 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.494925 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.555951 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.822937 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" event={"ID":"2084b413-8c86-4cc0-89ce-0dbfc4049e9b","Type":"ContainerStarted","Data":"04459e09511e8f218103b62e1cd093e2c1d10ce88498abbdd40b243c9345ef58"} Mar 21 05:05:24 crc kubenswrapper[4775]: I0321 05:05:24.958519 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.033316 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.159876 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.160984 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.163987 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.164176 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-hwff4" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.164187 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.171621 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.174205 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.182827 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.276245 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/76d205b7-bc2e-4dad-b513-457ff20d67e1-kolla-config\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.276299 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/76d205b7-bc2e-4dad-b513-457ff20d67e1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.276352 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/76d205b7-bc2e-4dad-b513-457ff20d67e1-config-data-default\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.276409 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.276431 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/76d205b7-bc2e-4dad-b513-457ff20d67e1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.276452 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76d205b7-bc2e-4dad-b513-457ff20d67e1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.276485 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76d205b7-bc2e-4dad-b513-457ff20d67e1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.276525 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wcnd\" (UniqueName: \"kubernetes.io/projected/76d205b7-bc2e-4dad-b513-457ff20d67e1-kube-api-access-8wcnd\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.379036 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76d205b7-bc2e-4dad-b513-457ff20d67e1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.379159 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wcnd\" (UniqueName: \"kubernetes.io/projected/76d205b7-bc2e-4dad-b513-457ff20d67e1-kube-api-access-8wcnd\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.379271 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/76d205b7-bc2e-4dad-b513-457ff20d67e1-kolla-config\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.379320 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/76d205b7-bc2e-4dad-b513-457ff20d67e1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.379398 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/76d205b7-bc2e-4dad-b513-457ff20d67e1-config-data-default\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.379478 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.379503 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/76d205b7-bc2e-4dad-b513-457ff20d67e1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.379524 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76d205b7-bc2e-4dad-b513-457ff20d67e1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.381133 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76d205b7-bc2e-4dad-b513-457ff20d67e1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.381183 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/76d205b7-bc2e-4dad-b513-457ff20d67e1-kolla-config\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.381435 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.381952 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/76d205b7-bc2e-4dad-b513-457ff20d67e1-config-data-default\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.382201 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/76d205b7-bc2e-4dad-b513-457ff20d67e1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.390492 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/76d205b7-bc2e-4dad-b513-457ff20d67e1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.394436 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76d205b7-bc2e-4dad-b513-457ff20d67e1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.399997 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wcnd\" (UniqueName: \"kubernetes.io/projected/76d205b7-bc2e-4dad-b513-457ff20d67e1-kube-api-access-8wcnd\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.454788 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"76d205b7-bc2e-4dad-b513-457ff20d67e1\") " pod="openstack/openstack-galera-0" Mar 21 05:05:25 crc kubenswrapper[4775]: I0321 05:05:25.486516 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.488436 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.490248 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.492021 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-sr8tt" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.492711 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.493318 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.500543 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.513336 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.595420 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bade9789-f227-44ab-b7fa-2173445cd381-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.595468 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bade9789-f227-44ab-b7fa-2173445cd381-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.595495 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.595533 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bade9789-f227-44ab-b7fa-2173445cd381-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.595550 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bade9789-f227-44ab-b7fa-2173445cd381-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.595583 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bade9789-f227-44ab-b7fa-2173445cd381-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.595622 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88pqb\" (UniqueName: \"kubernetes.io/projected/bade9789-f227-44ab-b7fa-2173445cd381-kube-api-access-88pqb\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.595660 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bade9789-f227-44ab-b7fa-2173445cd381-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.699968 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88pqb\" (UniqueName: \"kubernetes.io/projected/bade9789-f227-44ab-b7fa-2173445cd381-kube-api-access-88pqb\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.700023 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bade9789-f227-44ab-b7fa-2173445cd381-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.700063 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bade9789-f227-44ab-b7fa-2173445cd381-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.700082 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bade9789-f227-44ab-b7fa-2173445cd381-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.700106 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.700168 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bade9789-f227-44ab-b7fa-2173445cd381-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.700186 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bade9789-f227-44ab-b7fa-2173445cd381-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.700219 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bade9789-f227-44ab-b7fa-2173445cd381-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.700896 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bade9789-f227-44ab-b7fa-2173445cd381-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.701015 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.701256 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bade9789-f227-44ab-b7fa-2173445cd381-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.701775 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bade9789-f227-44ab-b7fa-2173445cd381-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.702082 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bade9789-f227-44ab-b7fa-2173445cd381-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.714312 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bade9789-f227-44ab-b7fa-2173445cd381-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.714468 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bade9789-f227-44ab-b7fa-2173445cd381-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.717953 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88pqb\" (UniqueName: \"kubernetes.io/projected/bade9789-f227-44ab-b7fa-2173445cd381-kube-api-access-88pqb\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.735186 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"bade9789-f227-44ab-b7fa-2173445cd381\") " pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.810794 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.883749 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.884992 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.888567 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.888843 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-wnhbq" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.889001 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Mar 21 05:05:26 crc kubenswrapper[4775]: I0321 05:05:26.895148 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 21 05:05:27 crc kubenswrapper[4775]: I0321 05:05:27.005614 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2a8305a2-5178-437f-a896-314b34fa595e-kolla-config\") pod \"memcached-0\" (UID: \"2a8305a2-5178-437f-a896-314b34fa595e\") " pod="openstack/memcached-0" Mar 21 05:05:27 crc kubenswrapper[4775]: I0321 05:05:27.005688 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pvwr\" (UniqueName: \"kubernetes.io/projected/2a8305a2-5178-437f-a896-314b34fa595e-kube-api-access-6pvwr\") pod \"memcached-0\" (UID: \"2a8305a2-5178-437f-a896-314b34fa595e\") " pod="openstack/memcached-0" Mar 21 05:05:27 crc kubenswrapper[4775]: I0321 05:05:27.005769 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a8305a2-5178-437f-a896-314b34fa595e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"2a8305a2-5178-437f-a896-314b34fa595e\") " pod="openstack/memcached-0" Mar 21 05:05:27 crc kubenswrapper[4775]: I0321 05:05:27.005930 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a8305a2-5178-437f-a896-314b34fa595e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"2a8305a2-5178-437f-a896-314b34fa595e\") " pod="openstack/memcached-0" Mar 21 05:05:27 crc kubenswrapper[4775]: I0321 05:05:27.005962 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2a8305a2-5178-437f-a896-314b34fa595e-config-data\") pod \"memcached-0\" (UID: \"2a8305a2-5178-437f-a896-314b34fa595e\") " pod="openstack/memcached-0" Mar 21 05:05:27 crc kubenswrapper[4775]: I0321 05:05:27.107234 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2a8305a2-5178-437f-a896-314b34fa595e-kolla-config\") pod \"memcached-0\" (UID: \"2a8305a2-5178-437f-a896-314b34fa595e\") " pod="openstack/memcached-0" Mar 21 05:05:27 crc kubenswrapper[4775]: I0321 05:05:27.107279 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pvwr\" (UniqueName: \"kubernetes.io/projected/2a8305a2-5178-437f-a896-314b34fa595e-kube-api-access-6pvwr\") pod \"memcached-0\" (UID: \"2a8305a2-5178-437f-a896-314b34fa595e\") " pod="openstack/memcached-0" Mar 21 05:05:27 crc kubenswrapper[4775]: I0321 05:05:27.107307 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a8305a2-5178-437f-a896-314b34fa595e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"2a8305a2-5178-437f-a896-314b34fa595e\") " pod="openstack/memcached-0" Mar 21 05:05:27 crc kubenswrapper[4775]: I0321 05:05:27.107325 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a8305a2-5178-437f-a896-314b34fa595e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"2a8305a2-5178-437f-a896-314b34fa595e\") " pod="openstack/memcached-0" Mar 21 05:05:27 crc kubenswrapper[4775]: I0321 05:05:27.107353 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2a8305a2-5178-437f-a896-314b34fa595e-config-data\") pod \"memcached-0\" (UID: \"2a8305a2-5178-437f-a896-314b34fa595e\") " pod="openstack/memcached-0" Mar 21 05:05:27 crc kubenswrapper[4775]: I0321 05:05:27.108211 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2a8305a2-5178-437f-a896-314b34fa595e-kolla-config\") pod \"memcached-0\" (UID: \"2a8305a2-5178-437f-a896-314b34fa595e\") " pod="openstack/memcached-0" Mar 21 05:05:27 crc kubenswrapper[4775]: I0321 05:05:27.108326 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2a8305a2-5178-437f-a896-314b34fa595e-config-data\") pod \"memcached-0\" (UID: \"2a8305a2-5178-437f-a896-314b34fa595e\") " pod="openstack/memcached-0" Mar 21 05:05:27 crc kubenswrapper[4775]: I0321 05:05:27.110360 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a8305a2-5178-437f-a896-314b34fa595e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"2a8305a2-5178-437f-a896-314b34fa595e\") " pod="openstack/memcached-0" Mar 21 05:05:27 crc kubenswrapper[4775]: I0321 05:05:27.111628 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a8305a2-5178-437f-a896-314b34fa595e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"2a8305a2-5178-437f-a896-314b34fa595e\") " pod="openstack/memcached-0" Mar 21 05:05:27 crc kubenswrapper[4775]: I0321 05:05:27.130225 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pvwr\" (UniqueName: \"kubernetes.io/projected/2a8305a2-5178-437f-a896-314b34fa595e-kube-api-access-6pvwr\") pod \"memcached-0\" (UID: \"2a8305a2-5178-437f-a896-314b34fa595e\") " pod="openstack/memcached-0" Mar 21 05:05:27 crc kubenswrapper[4775]: I0321 05:05:27.214985 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 21 05:05:27 crc kubenswrapper[4775]: W0321 05:05:27.440932 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod375fb8b7_b673_4fd7_ae51_5f82f33c196f.slice/crio-56ec7d33ff14d09de4f061f404a6116851871ee2ca592e6f5daea8ccedd576e8 WatchSource:0}: Error finding container 56ec7d33ff14d09de4f061f404a6116851871ee2ca592e6f5daea8ccedd576e8: Status 404 returned error can't find the container with id 56ec7d33ff14d09de4f061f404a6116851871ee2ca592e6f5daea8ccedd576e8 Mar 21 05:05:27 crc kubenswrapper[4775]: W0321 05:05:27.443324 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod839e915e_8197_48e9_8b69_56ac420a1eed.slice/crio-ef63fbfadda52109f2b3ecef0f45ed06c3a90225a2c4d5559784277a4c40c9dd WatchSource:0}: Error finding container ef63fbfadda52109f2b3ecef0f45ed06c3a90225a2c4d5559784277a4c40c9dd: Status 404 returned error can't find the container with id ef63fbfadda52109f2b3ecef0f45ed06c3a90225a2c4d5559784277a4c40c9dd Mar 21 05:05:27 crc kubenswrapper[4775]: I0321 05:05:27.852311 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"839e915e-8197-48e9-8b69-56ac420a1eed","Type":"ContainerStarted","Data":"ef63fbfadda52109f2b3ecef0f45ed06c3a90225a2c4d5559784277a4c40c9dd"} Mar 21 05:05:27 crc kubenswrapper[4775]: I0321 05:05:27.853424 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"375fb8b7-b673-4fd7-ae51-5f82f33c196f","Type":"ContainerStarted","Data":"56ec7d33ff14d09de4f061f404a6116851871ee2ca592e6f5daea8ccedd576e8"} Mar 21 05:05:29 crc kubenswrapper[4775]: I0321 05:05:29.255312 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 21 05:05:29 crc kubenswrapper[4775]: I0321 05:05:29.256435 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 21 05:05:29 crc kubenswrapper[4775]: I0321 05:05:29.258278 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-h2svw" Mar 21 05:05:29 crc kubenswrapper[4775]: I0321 05:05:29.265739 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 21 05:05:29 crc kubenswrapper[4775]: I0321 05:05:29.356944 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6zwf\" (UniqueName: \"kubernetes.io/projected/31578b15-f84b-4862-ae52-6720dac8f5e2-kube-api-access-l6zwf\") pod \"kube-state-metrics-0\" (UID: \"31578b15-f84b-4862-ae52-6720dac8f5e2\") " pod="openstack/kube-state-metrics-0" Mar 21 05:05:29 crc kubenswrapper[4775]: I0321 05:05:29.458149 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6zwf\" (UniqueName: \"kubernetes.io/projected/31578b15-f84b-4862-ae52-6720dac8f5e2-kube-api-access-l6zwf\") pod \"kube-state-metrics-0\" (UID: \"31578b15-f84b-4862-ae52-6720dac8f5e2\") " pod="openstack/kube-state-metrics-0" Mar 21 05:05:29 crc kubenswrapper[4775]: I0321 05:05:29.490764 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6zwf\" (UniqueName: \"kubernetes.io/projected/31578b15-f84b-4862-ae52-6720dac8f5e2-kube-api-access-l6zwf\") pod \"kube-state-metrics-0\" (UID: \"31578b15-f84b-4862-ae52-6720dac8f5e2\") " pod="openstack/kube-state-metrics-0" Mar 21 05:05:29 crc kubenswrapper[4775]: I0321 05:05:29.577553 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.475889 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-nmtjx"] Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.477152 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.479188 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.481481 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.494455 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-9cknj" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.495729 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-nmtjx"] Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.544948 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-frhpj"] Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.546417 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.561423 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-frhpj"] Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.622638 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a8e948c-2978-40c8-961b-1b010f7ea920-scripts\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.622717 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8a8e948c-2978-40c8-961b-1b010f7ea920-var-run\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.622922 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a8e948c-2978-40c8-961b-1b010f7ea920-ovn-controller-tls-certs\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.622991 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8a8e948c-2978-40c8-961b-1b010f7ea920-var-log-ovn\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.623021 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqkcl\" (UniqueName: \"kubernetes.io/projected/8a8e948c-2978-40c8-961b-1b010f7ea920-kube-api-access-wqkcl\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.623145 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8a8e948c-2978-40c8-961b-1b010f7ea920-var-run-ovn\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.623247 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8e948c-2978-40c8-961b-1b010f7ea920-combined-ca-bundle\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.724901 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8e948c-2978-40c8-961b-1b010f7ea920-combined-ca-bundle\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.724964 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/88a367f7-4951-4e7c-889f-d147676654f8-var-log\") pod \"ovn-controller-ovs-frhpj\" (UID: \"88a367f7-4951-4e7c-889f-d147676654f8\") " pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.725008 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a8e948c-2978-40c8-961b-1b010f7ea920-scripts\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.725054 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8a8e948c-2978-40c8-961b-1b010f7ea920-var-run\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.725091 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88a367f7-4951-4e7c-889f-d147676654f8-scripts\") pod \"ovn-controller-ovs-frhpj\" (UID: \"88a367f7-4951-4e7c-889f-d147676654f8\") " pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.725163 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a8e948c-2978-40c8-961b-1b010f7ea920-ovn-controller-tls-certs\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.725190 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m28vs\" (UniqueName: \"kubernetes.io/projected/88a367f7-4951-4e7c-889f-d147676654f8-kube-api-access-m28vs\") pod \"ovn-controller-ovs-frhpj\" (UID: \"88a367f7-4951-4e7c-889f-d147676654f8\") " pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.725212 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8a8e948c-2978-40c8-961b-1b010f7ea920-var-log-ovn\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.725293 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqkcl\" (UniqueName: \"kubernetes.io/projected/8a8e948c-2978-40c8-961b-1b010f7ea920-kube-api-access-wqkcl\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.725617 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/88a367f7-4951-4e7c-889f-d147676654f8-var-run\") pod \"ovn-controller-ovs-frhpj\" (UID: \"88a367f7-4951-4e7c-889f-d147676654f8\") " pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.725694 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/88a367f7-4951-4e7c-889f-d147676654f8-etc-ovs\") pod \"ovn-controller-ovs-frhpj\" (UID: \"88a367f7-4951-4e7c-889f-d147676654f8\") " pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.725718 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/88a367f7-4951-4e7c-889f-d147676654f8-var-lib\") pod \"ovn-controller-ovs-frhpj\" (UID: \"88a367f7-4951-4e7c-889f-d147676654f8\") " pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.725732 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8a8e948c-2978-40c8-961b-1b010f7ea920-var-log-ovn\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.725792 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8a8e948c-2978-40c8-961b-1b010f7ea920-var-run\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.725844 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8a8e948c-2978-40c8-961b-1b010f7ea920-var-run-ovn\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.726020 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8a8e948c-2978-40c8-961b-1b010f7ea920-var-run-ovn\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.728397 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a8e948c-2978-40c8-961b-1b010f7ea920-scripts\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.731657 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a8e948c-2978-40c8-961b-1b010f7ea920-ovn-controller-tls-certs\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.740620 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8e948c-2978-40c8-961b-1b010f7ea920-combined-ca-bundle\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.751463 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqkcl\" (UniqueName: \"kubernetes.io/projected/8a8e948c-2978-40c8-961b-1b010f7ea920-kube-api-access-wqkcl\") pod \"ovn-controller-nmtjx\" (UID: \"8a8e948c-2978-40c8-961b-1b010f7ea920\") " pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.797368 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.827288 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m28vs\" (UniqueName: \"kubernetes.io/projected/88a367f7-4951-4e7c-889f-d147676654f8-kube-api-access-m28vs\") pod \"ovn-controller-ovs-frhpj\" (UID: \"88a367f7-4951-4e7c-889f-d147676654f8\") " pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.827360 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/88a367f7-4951-4e7c-889f-d147676654f8-var-run\") pod \"ovn-controller-ovs-frhpj\" (UID: \"88a367f7-4951-4e7c-889f-d147676654f8\") " pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.827389 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/88a367f7-4951-4e7c-889f-d147676654f8-etc-ovs\") pod \"ovn-controller-ovs-frhpj\" (UID: \"88a367f7-4951-4e7c-889f-d147676654f8\") " pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.827412 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/88a367f7-4951-4e7c-889f-d147676654f8-var-lib\") pod \"ovn-controller-ovs-frhpj\" (UID: \"88a367f7-4951-4e7c-889f-d147676654f8\") " pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.827471 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/88a367f7-4951-4e7c-889f-d147676654f8-var-log\") pod \"ovn-controller-ovs-frhpj\" (UID: \"88a367f7-4951-4e7c-889f-d147676654f8\") " pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.827574 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88a367f7-4951-4e7c-889f-d147676654f8-scripts\") pod \"ovn-controller-ovs-frhpj\" (UID: \"88a367f7-4951-4e7c-889f-d147676654f8\") " pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.828295 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/88a367f7-4951-4e7c-889f-d147676654f8-var-log\") pod \"ovn-controller-ovs-frhpj\" (UID: \"88a367f7-4951-4e7c-889f-d147676654f8\") " pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.828821 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/88a367f7-4951-4e7c-889f-d147676654f8-var-run\") pod \"ovn-controller-ovs-frhpj\" (UID: \"88a367f7-4951-4e7c-889f-d147676654f8\") " pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.828949 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/88a367f7-4951-4e7c-889f-d147676654f8-etc-ovs\") pod \"ovn-controller-ovs-frhpj\" (UID: \"88a367f7-4951-4e7c-889f-d147676654f8\") " pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.829058 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/88a367f7-4951-4e7c-889f-d147676654f8-var-lib\") pod \"ovn-controller-ovs-frhpj\" (UID: \"88a367f7-4951-4e7c-889f-d147676654f8\") " pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.830774 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88a367f7-4951-4e7c-889f-d147676654f8-scripts\") pod \"ovn-controller-ovs-frhpj\" (UID: \"88a367f7-4951-4e7c-889f-d147676654f8\") " pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.846411 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m28vs\" (UniqueName: \"kubernetes.io/projected/88a367f7-4951-4e7c-889f-d147676654f8-kube-api-access-m28vs\") pod \"ovn-controller-ovs-frhpj\" (UID: \"88a367f7-4951-4e7c-889f-d147676654f8\") " pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:32 crc kubenswrapper[4775]: I0321 05:05:32.860787 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.330989 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.334573 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.337901 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.338013 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.338156 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-947qk" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.338204 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.341285 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.349024 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.437729 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-config\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.437787 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.437817 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.437859 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.437888 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.437916 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.438002 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.438039 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxdms\" (UniqueName: \"kubernetes.io/projected/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-kube-api-access-pxdms\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.540751 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.540819 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxdms\" (UniqueName: \"kubernetes.io/projected/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-kube-api-access-pxdms\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.540889 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-config\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.540914 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.540935 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.540966 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.540987 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.541011 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.542057 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.542354 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.542452 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-config\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.543164 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.544789 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.547047 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.547610 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.557763 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxdms\" (UniqueName: \"kubernetes.io/projected/b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e-kube-api-access-pxdms\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.562874 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e\") " pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:33 crc kubenswrapper[4775]: I0321 05:05:33.667485 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.024929 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.027273 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.030963 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.031266 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.038182 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.041580 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-n5ltt" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.089593 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.192390 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/93886182-fca2-42a9-a134-2243c7c7073d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.192621 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/93886182-fca2-42a9-a134-2243c7c7073d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.192722 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93886182-fca2-42a9-a134-2243c7c7073d-config\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.192805 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.192902 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/93886182-fca2-42a9-a134-2243c7c7073d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.192984 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/93886182-fca2-42a9-a134-2243c7c7073d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.193074 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf5qj\" (UniqueName: \"kubernetes.io/projected/93886182-fca2-42a9-a134-2243c7c7073d-kube-api-access-gf5qj\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.193195 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93886182-fca2-42a9-a134-2243c7c7073d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.294139 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/93886182-fca2-42a9-a134-2243c7c7073d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.294197 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93886182-fca2-42a9-a134-2243c7c7073d-config\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.294225 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.294244 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/93886182-fca2-42a9-a134-2243c7c7073d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.294262 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/93886182-fca2-42a9-a134-2243c7c7073d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.294293 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf5qj\" (UniqueName: \"kubernetes.io/projected/93886182-fca2-42a9-a134-2243c7c7073d-kube-api-access-gf5qj\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.294320 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93886182-fca2-42a9-a134-2243c7c7073d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.294351 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/93886182-fca2-42a9-a134-2243c7c7073d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.294748 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/93886182-fca2-42a9-a134-2243c7c7073d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.295901 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.296740 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/93886182-fca2-42a9-a134-2243c7c7073d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.297829 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93886182-fca2-42a9-a134-2243c7c7073d-config\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.302858 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/93886182-fca2-42a9-a134-2243c7c7073d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.303284 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93886182-fca2-42a9-a134-2243c7c7073d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.303465 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/93886182-fca2-42a9-a134-2243c7c7073d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.314425 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf5qj\" (UniqueName: \"kubernetes.io/projected/93886182-fca2-42a9-a134-2243c7c7073d-kube-api-access-gf5qj\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.333784 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"93886182-fca2-42a9-a134-2243c7c7073d\") " pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:36 crc kubenswrapper[4775]: I0321 05:05:36.344066 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:37 crc kubenswrapper[4775]: I0321 05:05:37.224157 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 21 05:05:37 crc kubenswrapper[4775]: E0321 05:05:37.615236 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 21 05:05:37 crc kubenswrapper[4775]: E0321 05:05:37.615687 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vlpch,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-d942p_openstack(62c4e9af-66f9-4f32-8c9a-30d4907a753c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 05:05:37 crc kubenswrapper[4775]: E0321 05:05:37.616900 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-d942p" podUID="62c4e9af-66f9-4f32-8c9a-30d4907a753c" Mar 21 05:05:38 crc kubenswrapper[4775]: W0321 05:05:38.626807 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a8305a2_5178_437f_a896_314b34fa595e.slice/crio-598009fbb019244e3b53d66fa63a72d4717d256096a6e5954a74b77181d01dc1 WatchSource:0}: Error finding container 598009fbb019244e3b53d66fa63a72d4717d256096a6e5954a74b77181d01dc1: Status 404 returned error can't find the container with id 598009fbb019244e3b53d66fa63a72d4717d256096a6e5954a74b77181d01dc1 Mar 21 05:05:38 crc kubenswrapper[4775]: E0321 05:05:38.660657 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 21 05:05:38 crc kubenswrapper[4775]: E0321 05:05:38.660833 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-74n7g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-bqfch_openstack(e8c11166-8b08-43d6-84e3-e5cbd0adbd8d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 05:05:38 crc kubenswrapper[4775]: E0321 05:05:38.662006 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-bqfch" podUID="e8c11166-8b08-43d6-84e3-e5cbd0adbd8d" Mar 21 05:05:38 crc kubenswrapper[4775]: I0321 05:05:38.746871 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-d942p" Mar 21 05:05:38 crc kubenswrapper[4775]: I0321 05:05:38.839252 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlpch\" (UniqueName: \"kubernetes.io/projected/62c4e9af-66f9-4f32-8c9a-30d4907a753c-kube-api-access-vlpch\") pod \"62c4e9af-66f9-4f32-8c9a-30d4907a753c\" (UID: \"62c4e9af-66f9-4f32-8c9a-30d4907a753c\") " Mar 21 05:05:38 crc kubenswrapper[4775]: I0321 05:05:38.840163 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62c4e9af-66f9-4f32-8c9a-30d4907a753c-config\") pod \"62c4e9af-66f9-4f32-8c9a-30d4907a753c\" (UID: \"62c4e9af-66f9-4f32-8c9a-30d4907a753c\") " Mar 21 05:05:38 crc kubenswrapper[4775]: I0321 05:05:38.844024 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62c4e9af-66f9-4f32-8c9a-30d4907a753c-config" (OuterVolumeSpecName: "config") pod "62c4e9af-66f9-4f32-8c9a-30d4907a753c" (UID: "62c4e9af-66f9-4f32-8c9a-30d4907a753c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:05:38 crc kubenswrapper[4775]: I0321 05:05:38.844268 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62c4e9af-66f9-4f32-8c9a-30d4907a753c-kube-api-access-vlpch" (OuterVolumeSpecName: "kube-api-access-vlpch") pod "62c4e9af-66f9-4f32-8c9a-30d4907a753c" (UID: "62c4e9af-66f9-4f32-8c9a-30d4907a753c"). InnerVolumeSpecName "kube-api-access-vlpch". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:05:38 crc kubenswrapper[4775]: I0321 05:05:38.942736 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlpch\" (UniqueName: \"kubernetes.io/projected/62c4e9af-66f9-4f32-8c9a-30d4907a753c-kube-api-access-vlpch\") on node \"crc\" DevicePath \"\"" Mar 21 05:05:38 crc kubenswrapper[4775]: I0321 05:05:38.943074 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62c4e9af-66f9-4f32-8c9a-30d4907a753c-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:05:38 crc kubenswrapper[4775]: I0321 05:05:38.980420 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-d942p" event={"ID":"62c4e9af-66f9-4f32-8c9a-30d4907a753c","Type":"ContainerDied","Data":"a875cb7a582da7fe33690731a6ec7fd7d012b80f676cca55452c73ba7b74cfd0"} Mar 21 05:05:38 crc kubenswrapper[4775]: I0321 05:05:38.980444 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-d942p" Mar 21 05:05:38 crc kubenswrapper[4775]: I0321 05:05:38.982202 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"2a8305a2-5178-437f-a896-314b34fa595e","Type":"ContainerStarted","Data":"598009fbb019244e3b53d66fa63a72d4717d256096a6e5954a74b77181d01dc1"} Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.073702 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-d942p"] Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.082199 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-d942p"] Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.088245 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 21 05:05:39 crc kubenswrapper[4775]: W0321 05:05:39.088746 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbade9789_f227_44ab_b7fa_2173445cd381.slice/crio-71a4bfd304371129a48c8ee069c1f253612b58ce94a9b582afab27fc7cfffa37 WatchSource:0}: Error finding container 71a4bfd304371129a48c8ee069c1f253612b58ce94a9b582afab27fc7cfffa37: Status 404 returned error can't find the container with id 71a4bfd304371129a48c8ee069c1f253612b58ce94a9b582afab27fc7cfffa37 Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.431445 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 21 05:05:39 crc kubenswrapper[4775]: W0321 05:05:39.440439 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31578b15_f84b_4862_ae52_6720dac8f5e2.slice/crio-95eba21f713625c8d0e5f46fb846b1edb0522a4501294e1679ea681e44f67781 WatchSource:0}: Error finding container 95eba21f713625c8d0e5f46fb846b1edb0522a4501294e1679ea681e44f67781: Status 404 returned error can't find the container with id 95eba21f713625c8d0e5f46fb846b1edb0522a4501294e1679ea681e44f67781 Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.445174 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-nmtjx"] Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.453767 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.491743 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-bqfch" Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.523728 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-frhpj"] Mar 21 05:05:39 crc kubenswrapper[4775]: W0321 05:05:39.524434 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88a367f7_4951_4e7c_889f_d147676654f8.slice/crio-c5fb9d3ae3c518dfedbe697bf5cdfbf1d08a6cfe78edab781073070def753b30 WatchSource:0}: Error finding container c5fb9d3ae3c518dfedbe697bf5cdfbf1d08a6cfe78edab781073070def753b30: Status 404 returned error can't find the container with id c5fb9d3ae3c518dfedbe697bf5cdfbf1d08a6cfe78edab781073070def753b30 Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.657301 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8c11166-8b08-43d6-84e3-e5cbd0adbd8d-config\") pod \"e8c11166-8b08-43d6-84e3-e5cbd0adbd8d\" (UID: \"e8c11166-8b08-43d6-84e3-e5cbd0adbd8d\") " Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.657462 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8c11166-8b08-43d6-84e3-e5cbd0adbd8d-dns-svc\") pod \"e8c11166-8b08-43d6-84e3-e5cbd0adbd8d\" (UID: \"e8c11166-8b08-43d6-84e3-e5cbd0adbd8d\") " Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.657569 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74n7g\" (UniqueName: \"kubernetes.io/projected/e8c11166-8b08-43d6-84e3-e5cbd0adbd8d-kube-api-access-74n7g\") pod \"e8c11166-8b08-43d6-84e3-e5cbd0adbd8d\" (UID: \"e8c11166-8b08-43d6-84e3-e5cbd0adbd8d\") " Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.658023 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8c11166-8b08-43d6-84e3-e5cbd0adbd8d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e8c11166-8b08-43d6-84e3-e5cbd0adbd8d" (UID: "e8c11166-8b08-43d6-84e3-e5cbd0adbd8d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.658064 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8c11166-8b08-43d6-84e3-e5cbd0adbd8d-config" (OuterVolumeSpecName: "config") pod "e8c11166-8b08-43d6-84e3-e5cbd0adbd8d" (UID: "e8c11166-8b08-43d6-84e3-e5cbd0adbd8d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.670902 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62c4e9af-66f9-4f32-8c9a-30d4907a753c" path="/var/lib/kubelet/pods/62c4e9af-66f9-4f32-8c9a-30d4907a753c/volumes" Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.759880 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8c11166-8b08-43d6-84e3-e5cbd0adbd8d-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.759918 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8c11166-8b08-43d6-84e3-e5cbd0adbd8d-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.796746 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8c11166-8b08-43d6-84e3-e5cbd0adbd8d-kube-api-access-74n7g" (OuterVolumeSpecName: "kube-api-access-74n7g") pod "e8c11166-8b08-43d6-84e3-e5cbd0adbd8d" (UID: "e8c11166-8b08-43d6-84e3-e5cbd0adbd8d"). InnerVolumeSpecName "kube-api-access-74n7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.861620 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74n7g\" (UniqueName: \"kubernetes.io/projected/e8c11166-8b08-43d6-84e3-e5cbd0adbd8d-kube-api-access-74n7g\") on node \"crc\" DevicePath \"\"" Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.990178 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"839e915e-8197-48e9-8b69-56ac420a1eed","Type":"ContainerStarted","Data":"5a7046cba7a8f5bae4cf5ebd270104b81a62d747ce978d3680ad2d2e0a16d243"} Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.992087 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"76d205b7-bc2e-4dad-b513-457ff20d67e1","Type":"ContainerStarted","Data":"0ee342247a2ef74219a0dfefa06608353417f7dec8acfc9478104d9cbd2b4c18"} Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.993905 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"375fb8b7-b673-4fd7-ae51-5f82f33c196f","Type":"ContainerStarted","Data":"1c4292485e2c0ee4f0c83f962fffc48a32f990f8105220ac94e44a4691b7ff1f"} Mar 21 05:05:39 crc kubenswrapper[4775]: I0321 05:05:39.994774 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nmtjx" event={"ID":"8a8e948c-2978-40c8-961b-1b010f7ea920","Type":"ContainerStarted","Data":"1c5e5cd13d3918f6a03a83aeea0df7177a08b6000e20544475fcf11597f71a2d"} Mar 21 05:05:40 crc kubenswrapper[4775]: I0321 05:05:39.998452 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-bqfch" event={"ID":"e8c11166-8b08-43d6-84e3-e5cbd0adbd8d","Type":"ContainerDied","Data":"59274326bf041a9ff442e3ad650874c430fb6d402b02a40d2d8ac6b554c247ca"} Mar 21 05:05:40 crc kubenswrapper[4775]: I0321 05:05:39.998515 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-bqfch" Mar 21 05:05:40 crc kubenswrapper[4775]: I0321 05:05:39.999659 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-frhpj" event={"ID":"88a367f7-4951-4e7c-889f-d147676654f8","Type":"ContainerStarted","Data":"c5fb9d3ae3c518dfedbe697bf5cdfbf1d08a6cfe78edab781073070def753b30"} Mar 21 05:05:40 crc kubenswrapper[4775]: I0321 05:05:40.000532 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bade9789-f227-44ab-b7fa-2173445cd381","Type":"ContainerStarted","Data":"71a4bfd304371129a48c8ee069c1f253612b58ce94a9b582afab27fc7cfffa37"} Mar 21 05:05:40 crc kubenswrapper[4775]: I0321 05:05:40.002085 4775 generic.go:334] "Generic (PLEG): container finished" podID="2084b413-8c86-4cc0-89ce-0dbfc4049e9b" containerID="f749dd6bac424eb0695958e8b0d01fc15e3e824b72ea62322c8b3aee65b4d200" exitCode=0 Mar 21 05:05:40 crc kubenswrapper[4775]: I0321 05:05:40.002176 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" event={"ID":"2084b413-8c86-4cc0-89ce-0dbfc4049e9b","Type":"ContainerDied","Data":"f749dd6bac424eb0695958e8b0d01fc15e3e824b72ea62322c8b3aee65b4d200"} Mar 21 05:05:40 crc kubenswrapper[4775]: I0321 05:05:40.003102 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"31578b15-f84b-4862-ae52-6720dac8f5e2","Type":"ContainerStarted","Data":"95eba21f713625c8d0e5f46fb846b1edb0522a4501294e1679ea681e44f67781"} Mar 21 05:05:40 crc kubenswrapper[4775]: I0321 05:05:40.005462 4775 generic.go:334] "Generic (PLEG): container finished" podID="6b9920a8-df59-4747-8f72-1f25d22517b3" containerID="e34c6feccb4c3623eb4d46c49fed2bc999053cbdc47d3e661349de5746c1ab7d" exitCode=0 Mar 21 05:05:40 crc kubenswrapper[4775]: I0321 05:05:40.005492 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" event={"ID":"6b9920a8-df59-4747-8f72-1f25d22517b3","Type":"ContainerDied","Data":"e34c6feccb4c3623eb4d46c49fed2bc999053cbdc47d3e661349de5746c1ab7d"} Mar 21 05:05:40 crc kubenswrapper[4775]: I0321 05:05:40.114193 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bqfch"] Mar 21 05:05:40 crc kubenswrapper[4775]: I0321 05:05:40.120182 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bqfch"] Mar 21 05:05:40 crc kubenswrapper[4775]: I0321 05:05:40.256094 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 21 05:05:40 crc kubenswrapper[4775]: I0321 05:05:40.376518 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 21 05:05:40 crc kubenswrapper[4775]: W0321 05:05:40.720797 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93886182_fca2_42a9_a134_2243c7c7073d.slice/crio-400536f01cd75d0d61183bbdf297be1b31f3ce3d729c063daac7126e63ada8d3 WatchSource:0}: Error finding container 400536f01cd75d0d61183bbdf297be1b31f3ce3d729c063daac7126e63ada8d3: Status 404 returned error can't find the container with id 400536f01cd75d0d61183bbdf297be1b31f3ce3d729c063daac7126e63ada8d3 Mar 21 05:05:40 crc kubenswrapper[4775]: W0321 05:05:40.723404 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4d6ed0c_fd91_4ed6_9a1c_0a31f78ed48e.slice/crio-de83a4eb6c823a9c0e42ab2a3e22cd64db36c616bd8f0506ecbc4b6065e74379 WatchSource:0}: Error finding container de83a4eb6c823a9c0e42ab2a3e22cd64db36c616bd8f0506ecbc4b6065e74379: Status 404 returned error can't find the container with id de83a4eb6c823a9c0e42ab2a3e22cd64db36c616bd8f0506ecbc4b6065e74379 Mar 21 05:05:41 crc kubenswrapper[4775]: I0321 05:05:41.016631 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e","Type":"ContainerStarted","Data":"de83a4eb6c823a9c0e42ab2a3e22cd64db36c616bd8f0506ecbc4b6065e74379"} Mar 21 05:05:41 crc kubenswrapper[4775]: I0321 05:05:41.018944 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"93886182-fca2-42a9-a134-2243c7c7073d","Type":"ContainerStarted","Data":"400536f01cd75d0d61183bbdf297be1b31f3ce3d729c063daac7126e63ada8d3"} Mar 21 05:05:41 crc kubenswrapper[4775]: I0321 05:05:41.674170 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8c11166-8b08-43d6-84e3-e5cbd0adbd8d" path="/var/lib/kubelet/pods/e8c11166-8b08-43d6-84e3-e5cbd0adbd8d/volumes" Mar 21 05:05:42 crc kubenswrapper[4775]: I0321 05:05:42.031143 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" event={"ID":"6b9920a8-df59-4747-8f72-1f25d22517b3","Type":"ContainerStarted","Data":"c5ba9d79e2eb52707c6820af305f99b1e2b4041e9f4aebd9fe6f0da5d03ca612"} Mar 21 05:05:42 crc kubenswrapper[4775]: I0321 05:05:42.031525 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" Mar 21 05:05:42 crc kubenswrapper[4775]: I0321 05:05:42.034838 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" event={"ID":"2084b413-8c86-4cc0-89ce-0dbfc4049e9b","Type":"ContainerStarted","Data":"9a3778eec58b097b77fc9e80249e34ac71c57ec91481179476b24785158c24e3"} Mar 21 05:05:42 crc kubenswrapper[4775]: I0321 05:05:42.034990 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" Mar 21 05:05:42 crc kubenswrapper[4775]: I0321 05:05:42.038217 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"2a8305a2-5178-437f-a896-314b34fa595e","Type":"ContainerStarted","Data":"d148febd920b4396b058dabe656656d420318d45ca05c3371173ea771fb88de7"} Mar 21 05:05:42 crc kubenswrapper[4775]: I0321 05:05:42.038316 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Mar 21 05:05:42 crc kubenswrapper[4775]: I0321 05:05:42.054086 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" podStartSLOduration=4.894380628 podStartE2EDuration="20.054063448s" podCreationTimestamp="2026-03-21 05:05:22 +0000 UTC" firstStartedPulling="2026-03-21 05:05:23.62537611 +0000 UTC m=+1076.601839734" lastFinishedPulling="2026-03-21 05:05:38.78505893 +0000 UTC m=+1091.761522554" observedRunningTime="2026-03-21 05:05:42.048885632 +0000 UTC m=+1095.025349266" watchObservedRunningTime="2026-03-21 05:05:42.054063448 +0000 UTC m=+1095.030527072" Mar 21 05:05:42 crc kubenswrapper[4775]: I0321 05:05:42.072012 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" podStartSLOduration=5.469301977 podStartE2EDuration="20.071992365s" podCreationTimestamp="2026-03-21 05:05:22 +0000 UTC" firstStartedPulling="2026-03-21 05:05:24.176916498 +0000 UTC m=+1077.153380122" lastFinishedPulling="2026-03-21 05:05:38.779606886 +0000 UTC m=+1091.756070510" observedRunningTime="2026-03-21 05:05:42.065038489 +0000 UTC m=+1095.041502113" watchObservedRunningTime="2026-03-21 05:05:42.071992365 +0000 UTC m=+1095.048455989" Mar 21 05:05:42 crc kubenswrapper[4775]: I0321 05:05:42.087788 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=13.930443911 podStartE2EDuration="16.087769652s" podCreationTimestamp="2026-03-21 05:05:26 +0000 UTC" firstStartedPulling="2026-03-21 05:05:38.645713179 +0000 UTC m=+1091.622176803" lastFinishedPulling="2026-03-21 05:05:40.80303892 +0000 UTC m=+1093.779502544" observedRunningTime="2026-03-21 05:05:42.084649913 +0000 UTC m=+1095.061113557" watchObservedRunningTime="2026-03-21 05:05:42.087769652 +0000 UTC m=+1095.064233276" Mar 21 05:05:47 crc kubenswrapper[4775]: I0321 05:05:47.097249 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"31578b15-f84b-4862-ae52-6720dac8f5e2","Type":"ContainerStarted","Data":"ebdac31be6ec007bb4ed96d43a4d78aab3ca618e5b3cc6ddb93c2810c9ec401a"} Mar 21 05:05:47 crc kubenswrapper[4775]: I0321 05:05:47.098017 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 21 05:05:47 crc kubenswrapper[4775]: I0321 05:05:47.099922 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nmtjx" event={"ID":"8a8e948c-2978-40c8-961b-1b010f7ea920","Type":"ContainerStarted","Data":"5903ae6dcf222706531c88d4b76781af7c8d83e54858da54fffe258cfd24a8d1"} Mar 21 05:05:47 crc kubenswrapper[4775]: I0321 05:05:47.100037 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-nmtjx" Mar 21 05:05:47 crc kubenswrapper[4775]: I0321 05:05:47.101408 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e","Type":"ContainerStarted","Data":"1b47eab34a6ed1ec2f0f84bb8a8895662f33297606bb0ccea9e7478bc25144cd"} Mar 21 05:05:47 crc kubenswrapper[4775]: I0321 05:05:47.102955 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-frhpj" event={"ID":"88a367f7-4951-4e7c-889f-d147676654f8","Type":"ContainerStarted","Data":"c0c441602e7099da4eb86364cb4742ce7c9df7ca08a5b6193a116451d7ebf21f"} Mar 21 05:05:47 crc kubenswrapper[4775]: I0321 05:05:47.104362 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bade9789-f227-44ab-b7fa-2173445cd381","Type":"ContainerStarted","Data":"48c166a356d78aa2e47a8f4d74ca291a5cf4a906e0b5939c25cc1658ceae0a52"} Mar 21 05:05:47 crc kubenswrapper[4775]: I0321 05:05:47.106071 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"93886182-fca2-42a9-a134-2243c7c7073d","Type":"ContainerStarted","Data":"07ba9f15899b01d878ebea12f580563712b0fb002c676c1b85c52d6bf53361c3"} Mar 21 05:05:47 crc kubenswrapper[4775]: I0321 05:05:47.108717 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"76d205b7-bc2e-4dad-b513-457ff20d67e1","Type":"ContainerStarted","Data":"96c9a92b11a655f0110f2a971188cfd82d21616cb9cd4cd24d771176f3603b12"} Mar 21 05:05:47 crc kubenswrapper[4775]: I0321 05:05:47.140978 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=11.074439585 podStartE2EDuration="18.140956757s" podCreationTimestamp="2026-03-21 05:05:29 +0000 UTC" firstStartedPulling="2026-03-21 05:05:39.443478401 +0000 UTC m=+1092.419942025" lastFinishedPulling="2026-03-21 05:05:46.509995573 +0000 UTC m=+1099.486459197" observedRunningTime="2026-03-21 05:05:47.115424825 +0000 UTC m=+1100.091888449" watchObservedRunningTime="2026-03-21 05:05:47.140956757 +0000 UTC m=+1100.117420381" Mar 21 05:05:47 crc kubenswrapper[4775]: I0321 05:05:47.143649 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-nmtjx" podStartSLOduration=8.489568095 podStartE2EDuration="15.143630943s" podCreationTimestamp="2026-03-21 05:05:32 +0000 UTC" firstStartedPulling="2026-03-21 05:05:39.463705923 +0000 UTC m=+1092.440169547" lastFinishedPulling="2026-03-21 05:05:46.117768771 +0000 UTC m=+1099.094232395" observedRunningTime="2026-03-21 05:05:47.134642299 +0000 UTC m=+1100.111105923" watchObservedRunningTime="2026-03-21 05:05:47.143630943 +0000 UTC m=+1100.120094567" Mar 21 05:05:47 crc kubenswrapper[4775]: I0321 05:05:47.216074 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Mar 21 05:05:48 crc kubenswrapper[4775]: I0321 05:05:48.086320 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" Mar 21 05:05:48 crc kubenswrapper[4775]: I0321 05:05:48.128719 4775 generic.go:334] "Generic (PLEG): container finished" podID="88a367f7-4951-4e7c-889f-d147676654f8" containerID="c0c441602e7099da4eb86364cb4742ce7c9df7ca08a5b6193a116451d7ebf21f" exitCode=0 Mar 21 05:05:48 crc kubenswrapper[4775]: I0321 05:05:48.130079 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-frhpj" event={"ID":"88a367f7-4951-4e7c-889f-d147676654f8","Type":"ContainerDied","Data":"c0c441602e7099da4eb86364cb4742ce7c9df7ca08a5b6193a116451d7ebf21f"} Mar 21 05:05:48 crc kubenswrapper[4775]: I0321 05:05:48.645073 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" Mar 21 05:05:48 crc kubenswrapper[4775]: I0321 05:05:48.695513 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-4rt6c"] Mar 21 05:05:48 crc kubenswrapper[4775]: I0321 05:05:48.695760 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" podUID="6b9920a8-df59-4747-8f72-1f25d22517b3" containerName="dnsmasq-dns" containerID="cri-o://c5ba9d79e2eb52707c6820af305f99b1e2b4041e9f4aebd9fe6f0da5d03ca612" gracePeriod=10 Mar 21 05:05:49 crc kubenswrapper[4775]: I0321 05:05:49.151384 4775 generic.go:334] "Generic (PLEG): container finished" podID="6b9920a8-df59-4747-8f72-1f25d22517b3" containerID="c5ba9d79e2eb52707c6820af305f99b1e2b4041e9f4aebd9fe6f0da5d03ca612" exitCode=0 Mar 21 05:05:49 crc kubenswrapper[4775]: I0321 05:05:49.151465 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" event={"ID":"6b9920a8-df59-4747-8f72-1f25d22517b3","Type":"ContainerDied","Data":"c5ba9d79e2eb52707c6820af305f99b1e2b4041e9f4aebd9fe6f0da5d03ca612"} Mar 21 05:05:49 crc kubenswrapper[4775]: I0321 05:05:49.154679 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-frhpj" event={"ID":"88a367f7-4951-4e7c-889f-d147676654f8","Type":"ContainerStarted","Data":"3fb00cdb5714dbc33207822ccac44aa4fc4bbc07c11261280f69e6610060af07"} Mar 21 05:05:49 crc kubenswrapper[4775]: I0321 05:05:49.697615 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fvhb4"] Mar 21 05:05:49 crc kubenswrapper[4775]: I0321 05:05:49.700653 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" Mar 21 05:05:49 crc kubenswrapper[4775]: I0321 05:05:49.717088 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fvhb4"] Mar 21 05:05:49 crc kubenswrapper[4775]: I0321 05:05:49.840313 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zcdx\" (UniqueName: \"kubernetes.io/projected/82af84d9-ff4c-4c51-9b1c-62f1c74b7de5-kube-api-access-8zcdx\") pod \"dnsmasq-dns-7cb5889db5-fvhb4\" (UID: \"82af84d9-ff4c-4c51-9b1c-62f1c74b7de5\") " pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" Mar 21 05:05:49 crc kubenswrapper[4775]: I0321 05:05:49.840536 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82af84d9-ff4c-4c51-9b1c-62f1c74b7de5-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-fvhb4\" (UID: \"82af84d9-ff4c-4c51-9b1c-62f1c74b7de5\") " pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" Mar 21 05:05:49 crc kubenswrapper[4775]: I0321 05:05:49.840595 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82af84d9-ff4c-4c51-9b1c-62f1c74b7de5-config\") pod \"dnsmasq-dns-7cb5889db5-fvhb4\" (UID: \"82af84d9-ff4c-4c51-9b1c-62f1c74b7de5\") " pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" Mar 21 05:05:49 crc kubenswrapper[4775]: I0321 05:05:49.941991 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zcdx\" (UniqueName: \"kubernetes.io/projected/82af84d9-ff4c-4c51-9b1c-62f1c74b7de5-kube-api-access-8zcdx\") pod \"dnsmasq-dns-7cb5889db5-fvhb4\" (UID: \"82af84d9-ff4c-4c51-9b1c-62f1c74b7de5\") " pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" Mar 21 05:05:49 crc kubenswrapper[4775]: I0321 05:05:49.942163 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82af84d9-ff4c-4c51-9b1c-62f1c74b7de5-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-fvhb4\" (UID: \"82af84d9-ff4c-4c51-9b1c-62f1c74b7de5\") " pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" Mar 21 05:05:49 crc kubenswrapper[4775]: I0321 05:05:49.943211 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82af84d9-ff4c-4c51-9b1c-62f1c74b7de5-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-fvhb4\" (UID: \"82af84d9-ff4c-4c51-9b1c-62f1c74b7de5\") " pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" Mar 21 05:05:49 crc kubenswrapper[4775]: I0321 05:05:49.943286 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82af84d9-ff4c-4c51-9b1c-62f1c74b7de5-config\") pod \"dnsmasq-dns-7cb5889db5-fvhb4\" (UID: \"82af84d9-ff4c-4c51-9b1c-62f1c74b7de5\") " pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" Mar 21 05:05:49 crc kubenswrapper[4775]: I0321 05:05:49.946038 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82af84d9-ff4c-4c51-9b1c-62f1c74b7de5-config\") pod \"dnsmasq-dns-7cb5889db5-fvhb4\" (UID: \"82af84d9-ff4c-4c51-9b1c-62f1c74b7de5\") " pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" Mar 21 05:05:49 crc kubenswrapper[4775]: I0321 05:05:49.995389 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zcdx\" (UniqueName: \"kubernetes.io/projected/82af84d9-ff4c-4c51-9b1c-62f1c74b7de5-kube-api-access-8zcdx\") pod \"dnsmasq-dns-7cb5889db5-fvhb4\" (UID: \"82af84d9-ff4c-4c51-9b1c-62f1c74b7de5\") " pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.059102 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.172903 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-frhpj" event={"ID":"88a367f7-4951-4e7c-889f-d147676654f8","Type":"ContainerStarted","Data":"90b39826220e47692b56ceb10eb4f54520e5729930cd0d0bc39ceb38bad7d13d"} Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.174245 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.174279 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.205648 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-frhpj" podStartSLOduration=12.47662192 podStartE2EDuration="18.205631277s" podCreationTimestamp="2026-03-21 05:05:32 +0000 UTC" firstStartedPulling="2026-03-21 05:05:39.526398336 +0000 UTC m=+1092.502861960" lastFinishedPulling="2026-03-21 05:05:45.255407693 +0000 UTC m=+1098.231871317" observedRunningTime="2026-03-21 05:05:50.197794945 +0000 UTC m=+1103.174258579" watchObservedRunningTime="2026-03-21 05:05:50.205631277 +0000 UTC m=+1103.182094901" Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.672137 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.757364 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zth55\" (UniqueName: \"kubernetes.io/projected/6b9920a8-df59-4747-8f72-1f25d22517b3-kube-api-access-zth55\") pod \"6b9920a8-df59-4747-8f72-1f25d22517b3\" (UID: \"6b9920a8-df59-4747-8f72-1f25d22517b3\") " Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.757456 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b9920a8-df59-4747-8f72-1f25d22517b3-config\") pod \"6b9920a8-df59-4747-8f72-1f25d22517b3\" (UID: \"6b9920a8-df59-4747-8f72-1f25d22517b3\") " Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.757504 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b9920a8-df59-4747-8f72-1f25d22517b3-dns-svc\") pod \"6b9920a8-df59-4747-8f72-1f25d22517b3\" (UID: \"6b9920a8-df59-4747-8f72-1f25d22517b3\") " Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.761920 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b9920a8-df59-4747-8f72-1f25d22517b3-kube-api-access-zth55" (OuterVolumeSpecName: "kube-api-access-zth55") pod "6b9920a8-df59-4747-8f72-1f25d22517b3" (UID: "6b9920a8-df59-4747-8f72-1f25d22517b3"). InnerVolumeSpecName "kube-api-access-zth55". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.792566 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b9920a8-df59-4747-8f72-1f25d22517b3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6b9920a8-df59-4747-8f72-1f25d22517b3" (UID: "6b9920a8-df59-4747-8f72-1f25d22517b3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.793431 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b9920a8-df59-4747-8f72-1f25d22517b3-config" (OuterVolumeSpecName: "config") pod "6b9920a8-df59-4747-8f72-1f25d22517b3" (UID: "6b9920a8-df59-4747-8f72-1f25d22517b3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.864511 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zth55\" (UniqueName: \"kubernetes.io/projected/6b9920a8-df59-4747-8f72-1f25d22517b3-kube-api-access-zth55\") on node \"crc\" DevicePath \"\"" Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.864547 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b9920a8-df59-4747-8f72-1f25d22517b3-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.864557 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b9920a8-df59-4747-8f72-1f25d22517b3-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.900266 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Mar 21 05:05:50 crc kubenswrapper[4775]: E0321 05:05:50.900742 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b9920a8-df59-4747-8f72-1f25d22517b3" containerName="init" Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.900760 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b9920a8-df59-4747-8f72-1f25d22517b3" containerName="init" Mar 21 05:05:50 crc kubenswrapper[4775]: E0321 05:05:50.900774 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b9920a8-df59-4747-8f72-1f25d22517b3" containerName="dnsmasq-dns" Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.900782 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b9920a8-df59-4747-8f72-1f25d22517b3" containerName="dnsmasq-dns" Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.901011 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b9920a8-df59-4747-8f72-1f25d22517b3" containerName="dnsmasq-dns" Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.906910 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.907049 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.908820 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.909018 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.909261 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-t4d72" Mar 21 05:05:50 crc kubenswrapper[4775]: I0321 05:05:50.909664 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.067396 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/8e93b938-c138-4cfc-a227-e1cd648ad59a-lock\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.067793 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/8e93b938-c138-4cfc-a227-e1cd648ad59a-cache\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.067876 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z78kf\" (UniqueName: \"kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-kube-api-access-z78kf\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.067906 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.067924 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e93b938-c138-4cfc-a227-e1cd648ad59a-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.067980 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.169329 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.169411 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/8e93b938-c138-4cfc-a227-e1cd648ad59a-lock\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.169468 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/8e93b938-c138-4cfc-a227-e1cd648ad59a-cache\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:51 crc kubenswrapper[4775]: E0321 05:05:51.169549 4775 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 21 05:05:51 crc kubenswrapper[4775]: E0321 05:05:51.169577 4775 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 21 05:05:51 crc kubenswrapper[4775]: E0321 05:05:51.169635 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift podName:8e93b938-c138-4cfc-a227-e1cd648ad59a nodeName:}" failed. No retries permitted until 2026-03-21 05:05:51.669614739 +0000 UTC m=+1104.646078413 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift") pod "swift-storage-0" (UID: "8e93b938-c138-4cfc-a227-e1cd648ad59a") : configmap "swift-ring-files" not found Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.169693 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z78kf\" (UniqueName: \"kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-kube-api-access-z78kf\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.169753 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.169774 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e93b938-c138-4cfc-a227-e1cd648ad59a-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.169893 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/8e93b938-c138-4cfc-a227-e1cd648ad59a-lock\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.170032 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/8e93b938-c138-4cfc-a227-e1cd648ad59a-cache\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.170136 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/swift-storage-0" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.173638 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e93b938-c138-4cfc-a227-e1cd648ad59a-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.185274 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" event={"ID":"6b9920a8-df59-4747-8f72-1f25d22517b3","Type":"ContainerDied","Data":"70d41f601713d43b2beb92b797cb5b68ffe82649d11afaeaa30ae3ca5021d995"} Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.185329 4775 scope.go:117] "RemoveContainer" containerID="c5ba9d79e2eb52707c6820af305f99b1e2b4041e9f4aebd9fe6f0da5d03ca612" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.185333 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-4rt6c" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.190727 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z78kf\" (UniqueName: \"kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-kube-api-access-z78kf\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.192942 4775 generic.go:334] "Generic (PLEG): container finished" podID="bade9789-f227-44ab-b7fa-2173445cd381" containerID="48c166a356d78aa2e47a8f4d74ca291a5cf4a906e0b5939c25cc1658ceae0a52" exitCode=0 Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.193000 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bade9789-f227-44ab-b7fa-2173445cd381","Type":"ContainerDied","Data":"48c166a356d78aa2e47a8f4d74ca291a5cf4a906e0b5939c25cc1658ceae0a52"} Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.199486 4775 generic.go:334] "Generic (PLEG): container finished" podID="76d205b7-bc2e-4dad-b513-457ff20d67e1" containerID="96c9a92b11a655f0110f2a971188cfd82d21616cb9cd4cd24d771176f3603b12" exitCode=0 Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.199582 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"76d205b7-bc2e-4dad-b513-457ff20d67e1","Type":"ContainerDied","Data":"96c9a92b11a655f0110f2a971188cfd82d21616cb9cd4cd24d771176f3603b12"} Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.204030 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.213237 4775 scope.go:117] "RemoveContainer" containerID="e34c6feccb4c3623eb4d46c49fed2bc999053cbdc47d3e661349de5746c1ab7d" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.273893 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-4rt6c"] Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.283743 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-4rt6c"] Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.325567 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fvhb4"] Mar 21 05:05:51 crc kubenswrapper[4775]: W0321 05:05:51.332980 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82af84d9_ff4c_4c51_9b1c_62f1c74b7de5.slice/crio-e02dcc00a6516fa9f13285030a0cc8babf7e17cec678d5d2033e37c7b73146b1 WatchSource:0}: Error finding container e02dcc00a6516fa9f13285030a0cc8babf7e17cec678d5d2033e37c7b73146b1: Status 404 returned error can't find the container with id e02dcc00a6516fa9f13285030a0cc8babf7e17cec678d5d2033e37c7b73146b1 Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.452414 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-kzll6"] Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.453540 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.455947 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.461099 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.461380 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.466723 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-kzll6"] Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.576636 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9e521c27-9d67-47bc-b6ac-74fabb543d3f-swiftconf\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.576706 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e521c27-9d67-47bc-b6ac-74fabb543d3f-scripts\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.576756 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9e521c27-9d67-47bc-b6ac-74fabb543d3f-ring-data-devices\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.576790 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e521c27-9d67-47bc-b6ac-74fabb543d3f-combined-ca-bundle\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.576854 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9e521c27-9d67-47bc-b6ac-74fabb543d3f-etc-swift\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.576877 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r55zr\" (UniqueName: \"kubernetes.io/projected/9e521c27-9d67-47bc-b6ac-74fabb543d3f-kube-api-access-r55zr\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.576934 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9e521c27-9d67-47bc-b6ac-74fabb543d3f-dispersionconf\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.670910 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b9920a8-df59-4747-8f72-1f25d22517b3" path="/var/lib/kubelet/pods/6b9920a8-df59-4747-8f72-1f25d22517b3/volumes" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.678457 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9e521c27-9d67-47bc-b6ac-74fabb543d3f-dispersionconf\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.678552 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9e521c27-9d67-47bc-b6ac-74fabb543d3f-swiftconf\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.678597 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e521c27-9d67-47bc-b6ac-74fabb543d3f-scripts\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.678636 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.678664 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9e521c27-9d67-47bc-b6ac-74fabb543d3f-ring-data-devices\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.678699 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e521c27-9d67-47bc-b6ac-74fabb543d3f-combined-ca-bundle\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.678769 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9e521c27-9d67-47bc-b6ac-74fabb543d3f-etc-swift\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.678811 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r55zr\" (UniqueName: \"kubernetes.io/projected/9e521c27-9d67-47bc-b6ac-74fabb543d3f-kube-api-access-r55zr\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: E0321 05:05:51.680181 4775 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 21 05:05:51 crc kubenswrapper[4775]: E0321 05:05:51.680222 4775 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 21 05:05:51 crc kubenswrapper[4775]: E0321 05:05:51.680285 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift podName:8e93b938-c138-4cfc-a227-e1cd648ad59a nodeName:}" failed. No retries permitted until 2026-03-21 05:05:52.68026257 +0000 UTC m=+1105.656726264 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift") pod "swift-storage-0" (UID: "8e93b938-c138-4cfc-a227-e1cd648ad59a") : configmap "swift-ring-files" not found Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.680647 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9e521c27-9d67-47bc-b6ac-74fabb543d3f-etc-swift\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.681239 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e521c27-9d67-47bc-b6ac-74fabb543d3f-scripts\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.681239 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9e521c27-9d67-47bc-b6ac-74fabb543d3f-ring-data-devices\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.683149 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9e521c27-9d67-47bc-b6ac-74fabb543d3f-swiftconf\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.683229 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9e521c27-9d67-47bc-b6ac-74fabb543d3f-dispersionconf\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.683711 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e521c27-9d67-47bc-b6ac-74fabb543d3f-combined-ca-bundle\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.694370 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r55zr\" (UniqueName: \"kubernetes.io/projected/9e521c27-9d67-47bc-b6ac-74fabb543d3f-kube-api-access-r55zr\") pod \"swift-ring-rebalance-kzll6\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:51 crc kubenswrapper[4775]: I0321 05:05:51.771263 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:05:52 crc kubenswrapper[4775]: I0321 05:05:52.214853 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"76d205b7-bc2e-4dad-b513-457ff20d67e1","Type":"ContainerStarted","Data":"d7da6ca9065a278564c662137d4b73030cfc2bcf91860ac5cc6099fc96832e02"} Mar 21 05:05:52 crc kubenswrapper[4775]: I0321 05:05:52.224467 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e","Type":"ContainerStarted","Data":"216f3fe3d5a6c74c85b4da2fb8fcc48f70ef16a89394a57a3d56a99fab9a5f54"} Mar 21 05:05:52 crc kubenswrapper[4775]: I0321 05:05:52.229867 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-kzll6"] Mar 21 05:05:52 crc kubenswrapper[4775]: I0321 05:05:52.232507 4775 generic.go:334] "Generic (PLEG): container finished" podID="82af84d9-ff4c-4c51-9b1c-62f1c74b7de5" containerID="d8e770cca74b17f5b344c07bee4bb08d1dc49f7e281098f9558394fc5dd7fa11" exitCode=0 Mar 21 05:05:52 crc kubenswrapper[4775]: I0321 05:05:52.232604 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" event={"ID":"82af84d9-ff4c-4c51-9b1c-62f1c74b7de5","Type":"ContainerDied","Data":"d8e770cca74b17f5b344c07bee4bb08d1dc49f7e281098f9558394fc5dd7fa11"} Mar 21 05:05:52 crc kubenswrapper[4775]: I0321 05:05:52.232640 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" event={"ID":"82af84d9-ff4c-4c51-9b1c-62f1c74b7de5","Type":"ContainerStarted","Data":"e02dcc00a6516fa9f13285030a0cc8babf7e17cec678d5d2033e37c7b73146b1"} Mar 21 05:05:52 crc kubenswrapper[4775]: I0321 05:05:52.234911 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bade9789-f227-44ab-b7fa-2173445cd381","Type":"ContainerStarted","Data":"532bc505798ddd24619287c620e7ed5ad9ed287bde700c8c2a1b7198767d03e1"} Mar 21 05:05:52 crc kubenswrapper[4775]: W0321 05:05:52.242682 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e521c27_9d67_47bc_b6ac_74fabb543d3f.slice/crio-eb20ad310e635a1adb6bbb9a38d57d53e3b88c4783ef5987e437714596390e99 WatchSource:0}: Error finding container eb20ad310e635a1adb6bbb9a38d57d53e3b88c4783ef5987e437714596390e99: Status 404 returned error can't find the container with id eb20ad310e635a1adb6bbb9a38d57d53e3b88c4783ef5987e437714596390e99 Mar 21 05:05:52 crc kubenswrapper[4775]: I0321 05:05:52.242812 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"93886182-fca2-42a9-a134-2243c7c7073d","Type":"ContainerStarted","Data":"ab36bf974063031155464ce2cec10f1d6d423f25084be3ae9828067d552ec991"} Mar 21 05:05:52 crc kubenswrapper[4775]: I0321 05:05:52.278205 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=22.721696571 podStartE2EDuration="28.278187119s" podCreationTimestamp="2026-03-21 05:05:24 +0000 UTC" firstStartedPulling="2026-03-21 05:05:39.464091214 +0000 UTC m=+1092.440554838" lastFinishedPulling="2026-03-21 05:05:45.020581762 +0000 UTC m=+1097.997045386" observedRunningTime="2026-03-21 05:05:52.244740143 +0000 UTC m=+1105.221203787" watchObservedRunningTime="2026-03-21 05:05:52.278187119 +0000 UTC m=+1105.254650743" Mar 21 05:05:52 crc kubenswrapper[4775]: I0321 05:05:52.281025 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=8.026247651 podStartE2EDuration="18.281014369s" podCreationTimestamp="2026-03-21 05:05:34 +0000 UTC" firstStartedPulling="2026-03-21 05:05:40.724806697 +0000 UTC m=+1093.701270321" lastFinishedPulling="2026-03-21 05:05:50.979573415 +0000 UTC m=+1103.956037039" observedRunningTime="2026-03-21 05:05:52.277542841 +0000 UTC m=+1105.254006465" watchObservedRunningTime="2026-03-21 05:05:52.281014369 +0000 UTC m=+1105.257477993" Mar 21 05:05:52 crc kubenswrapper[4775]: I0321 05:05:52.345835 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=10.1223692 podStartE2EDuration="20.345819132s" podCreationTimestamp="2026-03-21 05:05:32 +0000 UTC" firstStartedPulling="2026-03-21 05:05:40.728838491 +0000 UTC m=+1093.705302115" lastFinishedPulling="2026-03-21 05:05:50.952288423 +0000 UTC m=+1103.928752047" observedRunningTime="2026-03-21 05:05:52.34541472 +0000 UTC m=+1105.321878334" watchObservedRunningTime="2026-03-21 05:05:52.345819132 +0000 UTC m=+1105.322282756" Mar 21 05:05:52 crc kubenswrapper[4775]: I0321 05:05:52.373394 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=21.531953534 podStartE2EDuration="27.373375681s" podCreationTimestamp="2026-03-21 05:05:25 +0000 UTC" firstStartedPulling="2026-03-21 05:05:39.091499306 +0000 UTC m=+1092.067962930" lastFinishedPulling="2026-03-21 05:05:44.932921453 +0000 UTC m=+1097.909385077" observedRunningTime="2026-03-21 05:05:52.370843199 +0000 UTC m=+1105.347306843" watchObservedRunningTime="2026-03-21 05:05:52.373375681 +0000 UTC m=+1105.349839315" Mar 21 05:05:52 crc kubenswrapper[4775]: I0321 05:05:52.694431 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:52 crc kubenswrapper[4775]: E0321 05:05:52.694578 4775 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 21 05:05:52 crc kubenswrapper[4775]: E0321 05:05:52.694592 4775 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 21 05:05:52 crc kubenswrapper[4775]: E0321 05:05:52.694640 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift podName:8e93b938-c138-4cfc-a227-e1cd648ad59a nodeName:}" failed. No retries permitted until 2026-03-21 05:05:54.694626856 +0000 UTC m=+1107.671090480 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift") pod "swift-storage-0" (UID: "8e93b938-c138-4cfc-a227-e1cd648ad59a") : configmap "swift-ring-files" not found Mar 21 05:05:53 crc kubenswrapper[4775]: I0321 05:05:53.253963 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" event={"ID":"82af84d9-ff4c-4c51-9b1c-62f1c74b7de5","Type":"ContainerStarted","Data":"370a63757853aedcd7779b33f70117874887acb44e0ca8a8c4868f0ae046d389"} Mar 21 05:05:53 crc kubenswrapper[4775]: I0321 05:05:53.254335 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" Mar 21 05:05:53 crc kubenswrapper[4775]: I0321 05:05:53.257192 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-kzll6" event={"ID":"9e521c27-9d67-47bc-b6ac-74fabb543d3f","Type":"ContainerStarted","Data":"eb20ad310e635a1adb6bbb9a38d57d53e3b88c4783ef5987e437714596390e99"} Mar 21 05:05:53 crc kubenswrapper[4775]: I0321 05:05:53.285477 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" podStartSLOduration=4.285458245 podStartE2EDuration="4.285458245s" podCreationTimestamp="2026-03-21 05:05:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:05:53.282174792 +0000 UTC m=+1106.258638436" watchObservedRunningTime="2026-03-21 05:05:53.285458245 +0000 UTC m=+1106.261921869" Mar 21 05:05:53 crc kubenswrapper[4775]: I0321 05:05:53.672761 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:54 crc kubenswrapper[4775]: I0321 05:05:54.345451 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:54 crc kubenswrapper[4775]: I0321 05:05:54.390943 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:54 crc kubenswrapper[4775]: I0321 05:05:54.667780 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:54 crc kubenswrapper[4775]: I0321 05:05:54.708781 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:54 crc kubenswrapper[4775]: I0321 05:05:54.725439 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:54 crc kubenswrapper[4775]: E0321 05:05:54.725627 4775 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 21 05:05:54 crc kubenswrapper[4775]: E0321 05:05:54.725649 4775 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 21 05:05:54 crc kubenswrapper[4775]: E0321 05:05:54.725705 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift podName:8e93b938-c138-4cfc-a227-e1cd648ad59a nodeName:}" failed. No retries permitted until 2026-03-21 05:05:58.725685357 +0000 UTC m=+1111.702148981 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift") pod "swift-storage-0" (UID: "8e93b938-c138-4cfc-a227-e1cd648ad59a") : configmap "swift-ring-files" not found Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.281339 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.320775 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.323010 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.487124 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.488901 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fvhb4"] Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.489184 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" podUID="82af84d9-ff4c-4c51-9b1c-62f1c74b7de5" containerName="dnsmasq-dns" containerID="cri-o://370a63757853aedcd7779b33f70117874887acb44e0ca8a8c4868f0ae046d389" gracePeriod=10 Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.491820 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.515276 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-g8lds"] Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.518651 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-g8lds" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.521989 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.555546 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-g8lds"] Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.618276 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-ckh8b"] Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.619685 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-ckh8b" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.624155 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.638695 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-ckh8b"] Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.686978 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfb65104-69e6-4820-baab-3b0eea2a38d7-dns-svc\") pod \"dnsmasq-dns-57d65f699f-g8lds\" (UID: \"cfb65104-69e6-4820-baab-3b0eea2a38d7\") " pod="openstack/dnsmasq-dns-57d65f699f-g8lds" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.687042 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj9w2\" (UniqueName: \"kubernetes.io/projected/9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8-kube-api-access-gj9w2\") pod \"ovn-controller-metrics-ckh8b\" (UID: \"9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8\") " pod="openstack/ovn-controller-metrics-ckh8b" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.687086 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfb65104-69e6-4820-baab-3b0eea2a38d7-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-g8lds\" (UID: \"cfb65104-69e6-4820-baab-3b0eea2a38d7\") " pod="openstack/dnsmasq-dns-57d65f699f-g8lds" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.687107 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8-ovn-rundir\") pod \"ovn-controller-metrics-ckh8b\" (UID: \"9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8\") " pod="openstack/ovn-controller-metrics-ckh8b" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.687154 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8-combined-ca-bundle\") pod \"ovn-controller-metrics-ckh8b\" (UID: \"9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8\") " pod="openstack/ovn-controller-metrics-ckh8b" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.687188 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8-ovs-rundir\") pod \"ovn-controller-metrics-ckh8b\" (UID: \"9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8\") " pod="openstack/ovn-controller-metrics-ckh8b" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.687230 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfb65104-69e6-4820-baab-3b0eea2a38d7-config\") pod \"dnsmasq-dns-57d65f699f-g8lds\" (UID: \"cfb65104-69e6-4820-baab-3b0eea2a38d7\") " pod="openstack/dnsmasq-dns-57d65f699f-g8lds" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.687280 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8-config\") pod \"ovn-controller-metrics-ckh8b\" (UID: \"9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8\") " pod="openstack/ovn-controller-metrics-ckh8b" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.687308 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-ckh8b\" (UID: \"9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8\") " pod="openstack/ovn-controller-metrics-ckh8b" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.687339 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v4nh\" (UniqueName: \"kubernetes.io/projected/cfb65104-69e6-4820-baab-3b0eea2a38d7-kube-api-access-6v4nh\") pod \"dnsmasq-dns-57d65f699f-g8lds\" (UID: \"cfb65104-69e6-4820-baab-3b0eea2a38d7\") " pod="openstack/dnsmasq-dns-57d65f699f-g8lds" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.717789 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-g8lds"] Mar 21 05:05:55 crc kubenswrapper[4775]: E0321 05:05:55.720405 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-6v4nh ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-57d65f699f-g8lds" podUID="cfb65104-69e6-4820-baab-3b0eea2a38d7" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.755227 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-b4m98"] Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.756500 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.762838 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.764432 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-b4m98"] Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.789231 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8-config\") pod \"ovn-controller-metrics-ckh8b\" (UID: \"9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8\") " pod="openstack/ovn-controller-metrics-ckh8b" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.789290 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-ckh8b\" (UID: \"9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8\") " pod="openstack/ovn-controller-metrics-ckh8b" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.789328 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v4nh\" (UniqueName: \"kubernetes.io/projected/cfb65104-69e6-4820-baab-3b0eea2a38d7-kube-api-access-6v4nh\") pod \"dnsmasq-dns-57d65f699f-g8lds\" (UID: \"cfb65104-69e6-4820-baab-3b0eea2a38d7\") " pod="openstack/dnsmasq-dns-57d65f699f-g8lds" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.789358 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltd4s\" (UniqueName: \"kubernetes.io/projected/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-kube-api-access-ltd4s\") pod \"dnsmasq-dns-b8fbc5445-b4m98\" (UID: \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.789414 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-b4m98\" (UID: \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.789472 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfb65104-69e6-4820-baab-3b0eea2a38d7-dns-svc\") pod \"dnsmasq-dns-57d65f699f-g8lds\" (UID: \"cfb65104-69e6-4820-baab-3b0eea2a38d7\") " pod="openstack/dnsmasq-dns-57d65f699f-g8lds" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.789516 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj9w2\" (UniqueName: \"kubernetes.io/projected/9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8-kube-api-access-gj9w2\") pod \"ovn-controller-metrics-ckh8b\" (UID: \"9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8\") " pod="openstack/ovn-controller-metrics-ckh8b" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.789539 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-b4m98\" (UID: \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.789590 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfb65104-69e6-4820-baab-3b0eea2a38d7-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-g8lds\" (UID: \"cfb65104-69e6-4820-baab-3b0eea2a38d7\") " pod="openstack/dnsmasq-dns-57d65f699f-g8lds" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.789614 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8-ovn-rundir\") pod \"ovn-controller-metrics-ckh8b\" (UID: \"9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8\") " pod="openstack/ovn-controller-metrics-ckh8b" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.789653 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8-combined-ca-bundle\") pod \"ovn-controller-metrics-ckh8b\" (UID: \"9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8\") " pod="openstack/ovn-controller-metrics-ckh8b" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.789678 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-config\") pod \"dnsmasq-dns-b8fbc5445-b4m98\" (UID: \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.789864 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8-ovs-rundir\") pod \"ovn-controller-metrics-ckh8b\" (UID: \"9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8\") " pod="openstack/ovn-controller-metrics-ckh8b" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.789917 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-b4m98\" (UID: \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.789985 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfb65104-69e6-4820-baab-3b0eea2a38d7-config\") pod \"dnsmasq-dns-57d65f699f-g8lds\" (UID: \"cfb65104-69e6-4820-baab-3b0eea2a38d7\") " pod="openstack/dnsmasq-dns-57d65f699f-g8lds" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.790890 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfb65104-69e6-4820-baab-3b0eea2a38d7-config\") pod \"dnsmasq-dns-57d65f699f-g8lds\" (UID: \"cfb65104-69e6-4820-baab-3b0eea2a38d7\") " pod="openstack/dnsmasq-dns-57d65f699f-g8lds" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.792002 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8-config\") pod \"ovn-controller-metrics-ckh8b\" (UID: \"9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8\") " pod="openstack/ovn-controller-metrics-ckh8b" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.792550 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8-ovn-rundir\") pod \"ovn-controller-metrics-ckh8b\" (UID: \"9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8\") " pod="openstack/ovn-controller-metrics-ckh8b" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.793105 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfb65104-69e6-4820-baab-3b0eea2a38d7-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-g8lds\" (UID: \"cfb65104-69e6-4820-baab-3b0eea2a38d7\") " pod="openstack/dnsmasq-dns-57d65f699f-g8lds" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.793234 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8-ovs-rundir\") pod \"ovn-controller-metrics-ckh8b\" (UID: \"9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8\") " pod="openstack/ovn-controller-metrics-ckh8b" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.794031 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfb65104-69e6-4820-baab-3b0eea2a38d7-dns-svc\") pod \"dnsmasq-dns-57d65f699f-g8lds\" (UID: \"cfb65104-69e6-4820-baab-3b0eea2a38d7\") " pod="openstack/dnsmasq-dns-57d65f699f-g8lds" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.798922 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8-combined-ca-bundle\") pod \"ovn-controller-metrics-ckh8b\" (UID: \"9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8\") " pod="openstack/ovn-controller-metrics-ckh8b" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.801778 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.803211 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.805868 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-ckh8b\" (UID: \"9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8\") " pod="openstack/ovn-controller-metrics-ckh8b" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.806479 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-phfns" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.806647 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.806705 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.806923 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.838680 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj9w2\" (UniqueName: \"kubernetes.io/projected/9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8-kube-api-access-gj9w2\") pod \"ovn-controller-metrics-ckh8b\" (UID: \"9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8\") " pod="openstack/ovn-controller-metrics-ckh8b" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.838750 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.839295 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v4nh\" (UniqueName: \"kubernetes.io/projected/cfb65104-69e6-4820-baab-3b0eea2a38d7-kube-api-access-6v4nh\") pod \"dnsmasq-dns-57d65f699f-g8lds\" (UID: \"cfb65104-69e6-4820-baab-3b0eea2a38d7\") " pod="openstack/dnsmasq-dns-57d65f699f-g8lds" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.891241 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-config\") pod \"dnsmasq-dns-b8fbc5445-b4m98\" (UID: \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.891300 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/334a5c95-becc-4389-bb6f-50e5957cded6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.891322 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-b4m98\" (UID: \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.891347 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/334a5c95-becc-4389-bb6f-50e5957cded6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.891367 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/334a5c95-becc-4389-bb6f-50e5957cded6-scripts\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.891392 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/334a5c95-becc-4389-bb6f-50e5957cded6-config\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.891417 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltd4s\" (UniqueName: \"kubernetes.io/projected/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-kube-api-access-ltd4s\") pod \"dnsmasq-dns-b8fbc5445-b4m98\" (UID: \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.891454 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/334a5c95-becc-4389-bb6f-50e5957cded6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.891475 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-b4m98\" (UID: \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.891620 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xn4q\" (UniqueName: \"kubernetes.io/projected/334a5c95-becc-4389-bb6f-50e5957cded6-kube-api-access-5xn4q\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.891660 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-b4m98\" (UID: \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.891694 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/334a5c95-becc-4389-bb6f-50e5957cded6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.892864 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-config\") pod \"dnsmasq-dns-b8fbc5445-b4m98\" (UID: \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.892899 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-b4m98\" (UID: \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.892953 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-b4m98\" (UID: \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.893638 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-b4m98\" (UID: \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.908916 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltd4s\" (UniqueName: \"kubernetes.io/projected/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-kube-api-access-ltd4s\") pod \"dnsmasq-dns-b8fbc5445-b4m98\" (UID: \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.945391 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-ckh8b" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.992825 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/334a5c95-becc-4389-bb6f-50e5957cded6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.992891 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/334a5c95-becc-4389-bb6f-50e5957cded6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.992926 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/334a5c95-becc-4389-bb6f-50e5957cded6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.992952 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/334a5c95-becc-4389-bb6f-50e5957cded6-scripts\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.992980 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/334a5c95-becc-4389-bb6f-50e5957cded6-config\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.993015 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/334a5c95-becc-4389-bb6f-50e5957cded6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.993048 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xn4q\" (UniqueName: \"kubernetes.io/projected/334a5c95-becc-4389-bb6f-50e5957cded6-kube-api-access-5xn4q\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.993781 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/334a5c95-becc-4389-bb6f-50e5957cded6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.994317 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/334a5c95-becc-4389-bb6f-50e5957cded6-scripts\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.994670 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/334a5c95-becc-4389-bb6f-50e5957cded6-config\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:55 crc kubenswrapper[4775]: I0321 05:05:55.996753 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/334a5c95-becc-4389-bb6f-50e5957cded6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.000404 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/334a5c95-becc-4389-bb6f-50e5957cded6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.001797 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/334a5c95-becc-4389-bb6f-50e5957cded6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.011846 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xn4q\" (UniqueName: \"kubernetes.io/projected/334a5c95-becc-4389-bb6f-50e5957cded6-kube-api-access-5xn4q\") pod \"ovn-northd-0\" (UID: \"334a5c95-becc-4389-bb6f-50e5957cded6\") " pod="openstack/ovn-northd-0" Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.093442 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.203838 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.290044 4775 generic.go:334] "Generic (PLEG): container finished" podID="82af84d9-ff4c-4c51-9b1c-62f1c74b7de5" containerID="370a63757853aedcd7779b33f70117874887acb44e0ca8a8c4868f0ae046d389" exitCode=0 Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.290165 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" event={"ID":"82af84d9-ff4c-4c51-9b1c-62f1c74b7de5","Type":"ContainerDied","Data":"370a63757853aedcd7779b33f70117874887acb44e0ca8a8c4868f0ae046d389"} Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.290254 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-g8lds" Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.299256 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-g8lds" Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.399835 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfb65104-69e6-4820-baab-3b0eea2a38d7-config\") pod \"cfb65104-69e6-4820-baab-3b0eea2a38d7\" (UID: \"cfb65104-69e6-4820-baab-3b0eea2a38d7\") " Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.399946 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6v4nh\" (UniqueName: \"kubernetes.io/projected/cfb65104-69e6-4820-baab-3b0eea2a38d7-kube-api-access-6v4nh\") pod \"cfb65104-69e6-4820-baab-3b0eea2a38d7\" (UID: \"cfb65104-69e6-4820-baab-3b0eea2a38d7\") " Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.400027 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfb65104-69e6-4820-baab-3b0eea2a38d7-dns-svc\") pod \"cfb65104-69e6-4820-baab-3b0eea2a38d7\" (UID: \"cfb65104-69e6-4820-baab-3b0eea2a38d7\") " Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.400051 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfb65104-69e6-4820-baab-3b0eea2a38d7-ovsdbserver-nb\") pod \"cfb65104-69e6-4820-baab-3b0eea2a38d7\" (UID: \"cfb65104-69e6-4820-baab-3b0eea2a38d7\") " Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.400771 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfb65104-69e6-4820-baab-3b0eea2a38d7-config" (OuterVolumeSpecName: "config") pod "cfb65104-69e6-4820-baab-3b0eea2a38d7" (UID: "cfb65104-69e6-4820-baab-3b0eea2a38d7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.400784 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfb65104-69e6-4820-baab-3b0eea2a38d7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cfb65104-69e6-4820-baab-3b0eea2a38d7" (UID: "cfb65104-69e6-4820-baab-3b0eea2a38d7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.400809 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfb65104-69e6-4820-baab-3b0eea2a38d7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cfb65104-69e6-4820-baab-3b0eea2a38d7" (UID: "cfb65104-69e6-4820-baab-3b0eea2a38d7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.401417 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfb65104-69e6-4820-baab-3b0eea2a38d7-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.401438 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfb65104-69e6-4820-baab-3b0eea2a38d7-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.401447 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfb65104-69e6-4820-baab-3b0eea2a38d7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.404738 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfb65104-69e6-4820-baab-3b0eea2a38d7-kube-api-access-6v4nh" (OuterVolumeSpecName: "kube-api-access-6v4nh") pod "cfb65104-69e6-4820-baab-3b0eea2a38d7" (UID: "cfb65104-69e6-4820-baab-3b0eea2a38d7"). InnerVolumeSpecName "kube-api-access-6v4nh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.503423 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6v4nh\" (UniqueName: \"kubernetes.io/projected/cfb65104-69e6-4820-baab-3b0eea2a38d7-kube-api-access-6v4nh\") on node \"crc\" DevicePath \"\"" Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.810879 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:56 crc kubenswrapper[4775]: I0321 05:05:56.810932 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:57 crc kubenswrapper[4775]: I0321 05:05:57.308881 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-g8lds" Mar 21 05:05:57 crc kubenswrapper[4775]: I0321 05:05:57.376789 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-g8lds"] Mar 21 05:05:57 crc kubenswrapper[4775]: I0321 05:05:57.383867 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-g8lds"] Mar 21 05:05:57 crc kubenswrapper[4775]: I0321 05:05:57.675080 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfb65104-69e6-4820-baab-3b0eea2a38d7" path="/var/lib/kubelet/pods/cfb65104-69e6-4820-baab-3b0eea2a38d7/volumes" Mar 21 05:05:58 crc kubenswrapper[4775]: I0321 05:05:58.739690 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:05:58 crc kubenswrapper[4775]: E0321 05:05:58.739990 4775 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 21 05:05:58 crc kubenswrapper[4775]: E0321 05:05:58.740007 4775 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 21 05:05:58 crc kubenswrapper[4775]: E0321 05:05:58.740054 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift podName:8e93b938-c138-4cfc-a227-e1cd648ad59a nodeName:}" failed. No retries permitted until 2026-03-21 05:06:06.740038835 +0000 UTC m=+1119.716502469 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift") pod "swift-storage-0" (UID: "8e93b938-c138-4cfc-a227-e1cd648ad59a") : configmap "swift-ring-files" not found Mar 21 05:05:58 crc kubenswrapper[4775]: I0321 05:05:58.949431 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:59 crc kubenswrapper[4775]: I0321 05:05:59.034569 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Mar 21 05:05:59 crc kubenswrapper[4775]: I0321 05:05:59.583502 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.061612 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" podUID="82af84d9-ff4c-4c51-9b1c-62f1c74b7de5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.110:5353: connect: connection refused" Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.158558 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567826-89d6v"] Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.162318 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567826-89d6v" Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.165536 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.165816 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.170078 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.177848 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567826-89d6v"] Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.265101 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbtbc\" (UniqueName: \"kubernetes.io/projected/0c64d94e-917c-49b6-824b-0b2bdf9691ef-kube-api-access-wbtbc\") pod \"auto-csr-approver-29567826-89d6v\" (UID: \"0c64d94e-917c-49b6-824b-0b2bdf9691ef\") " pod="openshift-infra/auto-csr-approver-29567826-89d6v" Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.367212 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbtbc\" (UniqueName: \"kubernetes.io/projected/0c64d94e-917c-49b6-824b-0b2bdf9691ef-kube-api-access-wbtbc\") pod \"auto-csr-approver-29567826-89d6v\" (UID: \"0c64d94e-917c-49b6-824b-0b2bdf9691ef\") " pod="openshift-infra/auto-csr-approver-29567826-89d6v" Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.394656 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbtbc\" (UniqueName: \"kubernetes.io/projected/0c64d94e-917c-49b6-824b-0b2bdf9691ef-kube-api-access-wbtbc\") pod \"auto-csr-approver-29567826-89d6v\" (UID: \"0c64d94e-917c-49b6-824b-0b2bdf9691ef\") " pod="openshift-infra/auto-csr-approver-29567826-89d6v" Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.463979 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-b4m98"] Mar 21 05:06:00 crc kubenswrapper[4775]: W0321 05:06:00.468330 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b50afbd_31b8_40ff_bd7b_1ce5021e2837.slice/crio-2e772d1abd813f0f52a871bbd2aa2afc7b674d5369d7a3240dceb79e9ea9149f WatchSource:0}: Error finding container 2e772d1abd813f0f52a871bbd2aa2afc7b674d5369d7a3240dceb79e9ea9149f: Status 404 returned error can't find the container with id 2e772d1abd813f0f52a871bbd2aa2afc7b674d5369d7a3240dceb79e9ea9149f Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.470615 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.481646 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567826-89d6v" Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.552434 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.608745 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-ckh8b"] Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.673526 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82af84d9-ff4c-4c51-9b1c-62f1c74b7de5-dns-svc\") pod \"82af84d9-ff4c-4c51-9b1c-62f1c74b7de5\" (UID: \"82af84d9-ff4c-4c51-9b1c-62f1c74b7de5\") " Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.673892 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82af84d9-ff4c-4c51-9b1c-62f1c74b7de5-config\") pod \"82af84d9-ff4c-4c51-9b1c-62f1c74b7de5\" (UID: \"82af84d9-ff4c-4c51-9b1c-62f1c74b7de5\") " Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.674021 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zcdx\" (UniqueName: \"kubernetes.io/projected/82af84d9-ff4c-4c51-9b1c-62f1c74b7de5-kube-api-access-8zcdx\") pod \"82af84d9-ff4c-4c51-9b1c-62f1c74b7de5\" (UID: \"82af84d9-ff4c-4c51-9b1c-62f1c74b7de5\") " Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.680609 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82af84d9-ff4c-4c51-9b1c-62f1c74b7de5-kube-api-access-8zcdx" (OuterVolumeSpecName: "kube-api-access-8zcdx") pod "82af84d9-ff4c-4c51-9b1c-62f1c74b7de5" (UID: "82af84d9-ff4c-4c51-9b1c-62f1c74b7de5"). InnerVolumeSpecName "kube-api-access-8zcdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.756818 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82af84d9-ff4c-4c51-9b1c-62f1c74b7de5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "82af84d9-ff4c-4c51-9b1c-62f1c74b7de5" (UID: "82af84d9-ff4c-4c51-9b1c-62f1c74b7de5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.759716 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82af84d9-ff4c-4c51-9b1c-62f1c74b7de5-config" (OuterVolumeSpecName: "config") pod "82af84d9-ff4c-4c51-9b1c-62f1c74b7de5" (UID: "82af84d9-ff4c-4c51-9b1c-62f1c74b7de5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.776109 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zcdx\" (UniqueName: \"kubernetes.io/projected/82af84d9-ff4c-4c51-9b1c-62f1c74b7de5-kube-api-access-8zcdx\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.776154 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82af84d9-ff4c-4c51-9b1c-62f1c74b7de5-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.776164 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82af84d9-ff4c-4c51-9b1c-62f1c74b7de5-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:00 crc kubenswrapper[4775]: E0321 05:06:00.915989 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b50afbd_31b8_40ff_bd7b_1ce5021e2837.slice/crio-5d8a4462c4906d47c3863ff44ffe3802ba03cf186cd035ba4999f12c54eb2a8b.scope\": RecentStats: unable to find data in memory cache]" Mar 21 05:06:00 crc kubenswrapper[4775]: W0321 05:06:00.993964 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c64d94e_917c_49b6_824b_0b2bdf9691ef.slice/crio-70eac75c6e7292f7ac2d4dacfbd4b1082cd3e02756525e9a961e590c8b4fc563 WatchSource:0}: Error finding container 70eac75c6e7292f7ac2d4dacfbd4b1082cd3e02756525e9a961e590c8b4fc563: Status 404 returned error can't find the container with id 70eac75c6e7292f7ac2d4dacfbd4b1082cd3e02756525e9a961e590c8b4fc563 Mar 21 05:06:00 crc kubenswrapper[4775]: I0321 05:06:00.995263 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567826-89d6v"] Mar 21 05:06:01 crc kubenswrapper[4775]: I0321 05:06:01.340171 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" event={"ID":"82af84d9-ff4c-4c51-9b1c-62f1c74b7de5","Type":"ContainerDied","Data":"e02dcc00a6516fa9f13285030a0cc8babf7e17cec678d5d2033e37c7b73146b1"} Mar 21 05:06:01 crc kubenswrapper[4775]: I0321 05:06:01.340193 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-fvhb4" Mar 21 05:06:01 crc kubenswrapper[4775]: I0321 05:06:01.340442 4775 scope.go:117] "RemoveContainer" containerID="370a63757853aedcd7779b33f70117874887acb44e0ca8a8c4868f0ae046d389" Mar 21 05:06:01 crc kubenswrapper[4775]: I0321 05:06:01.341466 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567826-89d6v" event={"ID":"0c64d94e-917c-49b6-824b-0b2bdf9691ef","Type":"ContainerStarted","Data":"70eac75c6e7292f7ac2d4dacfbd4b1082cd3e02756525e9a961e590c8b4fc563"} Mar 21 05:06:01 crc kubenswrapper[4775]: I0321 05:06:01.343085 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-ckh8b" event={"ID":"9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8","Type":"ContainerStarted","Data":"1986b1ca4e6f2dfa09df73f55b760d00a5f511d18e0eba6fe1c0b623e4d746d3"} Mar 21 05:06:01 crc kubenswrapper[4775]: I0321 05:06:01.343141 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-ckh8b" event={"ID":"9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8","Type":"ContainerStarted","Data":"fef7f5ca2bc3d18af87381e34cb9dcefad0b96fc30f5bf7aace6bf0f4f58d596"} Mar 21 05:06:01 crc kubenswrapper[4775]: I0321 05:06:01.345341 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-kzll6" event={"ID":"9e521c27-9d67-47bc-b6ac-74fabb543d3f","Type":"ContainerStarted","Data":"449084f0366fca2d37734d9abfa93601028eb223fcd5ecd49cab997851f8fae3"} Mar 21 05:06:01 crc kubenswrapper[4775]: I0321 05:06:01.346638 4775 generic.go:334] "Generic (PLEG): container finished" podID="6b50afbd-31b8-40ff-bd7b-1ce5021e2837" containerID="5d8a4462c4906d47c3863ff44ffe3802ba03cf186cd035ba4999f12c54eb2a8b" exitCode=0 Mar 21 05:06:01 crc kubenswrapper[4775]: I0321 05:06:01.346677 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" event={"ID":"6b50afbd-31b8-40ff-bd7b-1ce5021e2837","Type":"ContainerDied","Data":"5d8a4462c4906d47c3863ff44ffe3802ba03cf186cd035ba4999f12c54eb2a8b"} Mar 21 05:06:01 crc kubenswrapper[4775]: I0321 05:06:01.346691 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" event={"ID":"6b50afbd-31b8-40ff-bd7b-1ce5021e2837","Type":"ContainerStarted","Data":"2e772d1abd813f0f52a871bbd2aa2afc7b674d5369d7a3240dceb79e9ea9149f"} Mar 21 05:06:01 crc kubenswrapper[4775]: I0321 05:06:01.349016 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"334a5c95-becc-4389-bb6f-50e5957cded6","Type":"ContainerStarted","Data":"7693c5ea17ce78ac391cf46e868c97b34eed5f12803f81339f6dbbb8948bc8c1"} Mar 21 05:06:01 crc kubenswrapper[4775]: I0321 05:06:01.362965 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-ckh8b" podStartSLOduration=6.362948562 podStartE2EDuration="6.362948562s" podCreationTimestamp="2026-03-21 05:05:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:06:01.362264503 +0000 UTC m=+1114.338728117" watchObservedRunningTime="2026-03-21 05:06:01.362948562 +0000 UTC m=+1114.339412186" Mar 21 05:06:01 crc kubenswrapper[4775]: I0321 05:06:01.431184 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-kzll6" podStartSLOduration=2.1255120610000002 podStartE2EDuration="10.431162151s" podCreationTimestamp="2026-03-21 05:05:51 +0000 UTC" firstStartedPulling="2026-03-21 05:05:52.258537513 +0000 UTC m=+1105.235001137" lastFinishedPulling="2026-03-21 05:06:00.564187603 +0000 UTC m=+1113.540651227" observedRunningTime="2026-03-21 05:06:01.423990018 +0000 UTC m=+1114.400453642" watchObservedRunningTime="2026-03-21 05:06:01.431162151 +0000 UTC m=+1114.407625775" Mar 21 05:06:01 crc kubenswrapper[4775]: I0321 05:06:01.441566 4775 scope.go:117] "RemoveContainer" containerID="d8e770cca74b17f5b344c07bee4bb08d1dc49f7e281098f9558394fc5dd7fa11" Mar 21 05:06:01 crc kubenswrapper[4775]: I0321 05:06:01.451589 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fvhb4"] Mar 21 05:06:01 crc kubenswrapper[4775]: I0321 05:06:01.462654 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fvhb4"] Mar 21 05:06:01 crc kubenswrapper[4775]: I0321 05:06:01.660765 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Mar 21 05:06:01 crc kubenswrapper[4775]: I0321 05:06:01.674287 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82af84d9-ff4c-4c51-9b1c-62f1c74b7de5" path="/var/lib/kubelet/pods/82af84d9-ff4c-4c51-9b1c-62f1c74b7de5/volumes" Mar 21 05:06:01 crc kubenswrapper[4775]: I0321 05:06:01.736354 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Mar 21 05:06:02 crc kubenswrapper[4775]: I0321 05:06:02.368877 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" event={"ID":"6b50afbd-31b8-40ff-bd7b-1ce5021e2837","Type":"ContainerStarted","Data":"6c721bd1cf19d6ad1fc2eb055e74ccf9afea13852ad17eeb02e87ea598c42465"} Mar 21 05:06:02 crc kubenswrapper[4775]: I0321 05:06:02.369039 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:06:02 crc kubenswrapper[4775]: I0321 05:06:02.370392 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"334a5c95-becc-4389-bb6f-50e5957cded6","Type":"ContainerStarted","Data":"9c5152fe0933c9b96daebbc84ec37854c159baf333677ab793d9a9af111e7842"} Mar 21 05:06:02 crc kubenswrapper[4775]: I0321 05:06:02.370409 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"334a5c95-becc-4389-bb6f-50e5957cded6","Type":"ContainerStarted","Data":"d667e95ad37e5a372c98b7d3347afd11c23f24668a2941bd2a8ef2bcab231965"} Mar 21 05:06:02 crc kubenswrapper[4775]: I0321 05:06:02.370667 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Mar 21 05:06:02 crc kubenswrapper[4775]: I0321 05:06:02.373579 4775 generic.go:334] "Generic (PLEG): container finished" podID="0c64d94e-917c-49b6-824b-0b2bdf9691ef" containerID="2acf177da0f2ed7ac0606a427af701da980e32ea57fc8b3636ddd64f2a8a5536" exitCode=0 Mar 21 05:06:02 crc kubenswrapper[4775]: I0321 05:06:02.373682 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567826-89d6v" event={"ID":"0c64d94e-917c-49b6-824b-0b2bdf9691ef","Type":"ContainerDied","Data":"2acf177da0f2ed7ac0606a427af701da980e32ea57fc8b3636ddd64f2a8a5536"} Mar 21 05:06:02 crc kubenswrapper[4775]: I0321 05:06:02.389461 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" podStartSLOduration=7.389399521 podStartE2EDuration="7.389399521s" podCreationTimestamp="2026-03-21 05:05:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:06:02.385542431 +0000 UTC m=+1115.362006045" watchObservedRunningTime="2026-03-21 05:06:02.389399521 +0000 UTC m=+1115.365863145" Mar 21 05:06:02 crc kubenswrapper[4775]: I0321 05:06:02.411850 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=6.392096286 podStartE2EDuration="7.411828115s" podCreationTimestamp="2026-03-21 05:05:55 +0000 UTC" firstStartedPulling="2026-03-21 05:06:00.475676899 +0000 UTC m=+1113.452140523" lastFinishedPulling="2026-03-21 05:06:01.495408728 +0000 UTC m=+1114.471872352" observedRunningTime="2026-03-21 05:06:02.406763712 +0000 UTC m=+1115.383227336" watchObservedRunningTime="2026-03-21 05:06:02.411828115 +0000 UTC m=+1115.388291739" Mar 21 05:06:03 crc kubenswrapper[4775]: I0321 05:06:03.945725 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567826-89d6v" Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.058517 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbtbc\" (UniqueName: \"kubernetes.io/projected/0c64d94e-917c-49b6-824b-0b2bdf9691ef-kube-api-access-wbtbc\") pod \"0c64d94e-917c-49b6-824b-0b2bdf9691ef\" (UID: \"0c64d94e-917c-49b6-824b-0b2bdf9691ef\") " Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.064796 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c64d94e-917c-49b6-824b-0b2bdf9691ef-kube-api-access-wbtbc" (OuterVolumeSpecName: "kube-api-access-wbtbc") pod "0c64d94e-917c-49b6-824b-0b2bdf9691ef" (UID: "0c64d94e-917c-49b6-824b-0b2bdf9691ef"). InnerVolumeSpecName "kube-api-access-wbtbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.160887 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbtbc\" (UniqueName: \"kubernetes.io/projected/0c64d94e-917c-49b6-824b-0b2bdf9691ef-kube-api-access-wbtbc\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.162738 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-tq6zw"] Mar 21 05:06:04 crc kubenswrapper[4775]: E0321 05:06:04.163104 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c64d94e-917c-49b6-824b-0b2bdf9691ef" containerName="oc" Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.163143 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c64d94e-917c-49b6-824b-0b2bdf9691ef" containerName="oc" Mar 21 05:06:04 crc kubenswrapper[4775]: E0321 05:06:04.163170 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82af84d9-ff4c-4c51-9b1c-62f1c74b7de5" containerName="init" Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.163178 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="82af84d9-ff4c-4c51-9b1c-62f1c74b7de5" containerName="init" Mar 21 05:06:04 crc kubenswrapper[4775]: E0321 05:06:04.163204 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82af84d9-ff4c-4c51-9b1c-62f1c74b7de5" containerName="dnsmasq-dns" Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.163212 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="82af84d9-ff4c-4c51-9b1c-62f1c74b7de5" containerName="dnsmasq-dns" Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.163401 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="82af84d9-ff4c-4c51-9b1c-62f1c74b7de5" containerName="dnsmasq-dns" Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.163423 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c64d94e-917c-49b6-824b-0b2bdf9691ef" containerName="oc" Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.163949 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-tq6zw" Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.165590 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.180858 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-tq6zw"] Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.262221 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af1531c7-e7a2-45cd-a5b9-9678c4808315-operator-scripts\") pod \"root-account-create-update-tq6zw\" (UID: \"af1531c7-e7a2-45cd-a5b9-9678c4808315\") " pod="openstack/root-account-create-update-tq6zw" Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.262285 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rn42\" (UniqueName: \"kubernetes.io/projected/af1531c7-e7a2-45cd-a5b9-9678c4808315-kube-api-access-4rn42\") pod \"root-account-create-update-tq6zw\" (UID: \"af1531c7-e7a2-45cd-a5b9-9678c4808315\") " pod="openstack/root-account-create-update-tq6zw" Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.364331 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af1531c7-e7a2-45cd-a5b9-9678c4808315-operator-scripts\") pod \"root-account-create-update-tq6zw\" (UID: \"af1531c7-e7a2-45cd-a5b9-9678c4808315\") " pod="openstack/root-account-create-update-tq6zw" Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.364412 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rn42\" (UniqueName: \"kubernetes.io/projected/af1531c7-e7a2-45cd-a5b9-9678c4808315-kube-api-access-4rn42\") pod \"root-account-create-update-tq6zw\" (UID: \"af1531c7-e7a2-45cd-a5b9-9678c4808315\") " pod="openstack/root-account-create-update-tq6zw" Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.364998 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af1531c7-e7a2-45cd-a5b9-9678c4808315-operator-scripts\") pod \"root-account-create-update-tq6zw\" (UID: \"af1531c7-e7a2-45cd-a5b9-9678c4808315\") " pod="openstack/root-account-create-update-tq6zw" Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.385009 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rn42\" (UniqueName: \"kubernetes.io/projected/af1531c7-e7a2-45cd-a5b9-9678c4808315-kube-api-access-4rn42\") pod \"root-account-create-update-tq6zw\" (UID: \"af1531c7-e7a2-45cd-a5b9-9678c4808315\") " pod="openstack/root-account-create-update-tq6zw" Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.484140 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-tq6zw" Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.581136 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567826-89d6v" event={"ID":"0c64d94e-917c-49b6-824b-0b2bdf9691ef","Type":"ContainerDied","Data":"70eac75c6e7292f7ac2d4dacfbd4b1082cd3e02756525e9a961e590c8b4fc563"} Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.581173 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567826-89d6v" Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.581183 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70eac75c6e7292f7ac2d4dacfbd4b1082cd3e02756525e9a961e590c8b4fc563" Mar 21 05:06:04 crc kubenswrapper[4775]: I0321 05:06:04.949464 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-tq6zw"] Mar 21 05:06:05 crc kubenswrapper[4775]: I0321 05:06:05.018678 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567820-cngbx"] Mar 21 05:06:05 crc kubenswrapper[4775]: I0321 05:06:05.027913 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567820-cngbx"] Mar 21 05:06:05 crc kubenswrapper[4775]: I0321 05:06:05.592532 4775 generic.go:334] "Generic (PLEG): container finished" podID="af1531c7-e7a2-45cd-a5b9-9678c4808315" containerID="8177bcef8f42975589ec3469527727473adafaa3cc6210ffa908111b0912b0d5" exitCode=0 Mar 21 05:06:05 crc kubenswrapper[4775]: I0321 05:06:05.592740 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-tq6zw" event={"ID":"af1531c7-e7a2-45cd-a5b9-9678c4808315","Type":"ContainerDied","Data":"8177bcef8f42975589ec3469527727473adafaa3cc6210ffa908111b0912b0d5"} Mar 21 05:06:05 crc kubenswrapper[4775]: I0321 05:06:05.592821 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-tq6zw" event={"ID":"af1531c7-e7a2-45cd-a5b9-9678c4808315","Type":"ContainerStarted","Data":"490b9754f17abc231dc24b2f80cf9533e36998bd95e31c2f91871cb4975c2c6c"} Mar 21 05:06:05 crc kubenswrapper[4775]: I0321 05:06:05.677841 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3" path="/var/lib/kubelet/pods/1a4faca2-a8cb-41b5-a0cf-44be1cea8fd3/volumes" Mar 21 05:06:06 crc kubenswrapper[4775]: I0321 05:06:06.096772 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:06:06 crc kubenswrapper[4775]: I0321 05:06:06.163549 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mn5xb"] Mar 21 05:06:06 crc kubenswrapper[4775]: I0321 05:06:06.163876 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" podUID="2084b413-8c86-4cc0-89ce-0dbfc4049e9b" containerName="dnsmasq-dns" containerID="cri-o://9a3778eec58b097b77fc9e80249e34ac71c57ec91481179476b24785158c24e3" gracePeriod=10 Mar 21 05:06:06 crc kubenswrapper[4775]: I0321 05:06:06.601462 4775 generic.go:334] "Generic (PLEG): container finished" podID="2084b413-8c86-4cc0-89ce-0dbfc4049e9b" containerID="9a3778eec58b097b77fc9e80249e34ac71c57ec91481179476b24785158c24e3" exitCode=0 Mar 21 05:06:06 crc kubenswrapper[4775]: I0321 05:06:06.601558 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" event={"ID":"2084b413-8c86-4cc0-89ce-0dbfc4049e9b","Type":"ContainerDied","Data":"9a3778eec58b097b77fc9e80249e34ac71c57ec91481179476b24785158c24e3"} Mar 21 05:06:06 crc kubenswrapper[4775]: I0321 05:06:06.601985 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" event={"ID":"2084b413-8c86-4cc0-89ce-0dbfc4049e9b","Type":"ContainerDied","Data":"04459e09511e8f218103b62e1cd093e2c1d10ce88498abbdd40b243c9345ef58"} Mar 21 05:06:06 crc kubenswrapper[4775]: I0321 05:06:06.602003 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04459e09511e8f218103b62e1cd093e2c1d10ce88498abbdd40b243c9345ef58" Mar 21 05:06:06 crc kubenswrapper[4775]: I0321 05:06:06.640020 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" Mar 21 05:06:06 crc kubenswrapper[4775]: I0321 05:06:06.712375 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f77fw\" (UniqueName: \"kubernetes.io/projected/2084b413-8c86-4cc0-89ce-0dbfc4049e9b-kube-api-access-f77fw\") pod \"2084b413-8c86-4cc0-89ce-0dbfc4049e9b\" (UID: \"2084b413-8c86-4cc0-89ce-0dbfc4049e9b\") " Mar 21 05:06:06 crc kubenswrapper[4775]: I0321 05:06:06.712477 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2084b413-8c86-4cc0-89ce-0dbfc4049e9b-config\") pod \"2084b413-8c86-4cc0-89ce-0dbfc4049e9b\" (UID: \"2084b413-8c86-4cc0-89ce-0dbfc4049e9b\") " Mar 21 05:06:06 crc kubenswrapper[4775]: I0321 05:06:06.713488 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2084b413-8c86-4cc0-89ce-0dbfc4049e9b-dns-svc\") pod \"2084b413-8c86-4cc0-89ce-0dbfc4049e9b\" (UID: \"2084b413-8c86-4cc0-89ce-0dbfc4049e9b\") " Mar 21 05:06:06 crc kubenswrapper[4775]: I0321 05:06:06.730604 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2084b413-8c86-4cc0-89ce-0dbfc4049e9b-kube-api-access-f77fw" (OuterVolumeSpecName: "kube-api-access-f77fw") pod "2084b413-8c86-4cc0-89ce-0dbfc4049e9b" (UID: "2084b413-8c86-4cc0-89ce-0dbfc4049e9b"). InnerVolumeSpecName "kube-api-access-f77fw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:06 crc kubenswrapper[4775]: I0321 05:06:06.753707 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2084b413-8c86-4cc0-89ce-0dbfc4049e9b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2084b413-8c86-4cc0-89ce-0dbfc4049e9b" (UID: "2084b413-8c86-4cc0-89ce-0dbfc4049e9b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:06 crc kubenswrapper[4775]: I0321 05:06:06.785127 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2084b413-8c86-4cc0-89ce-0dbfc4049e9b-config" (OuterVolumeSpecName: "config") pod "2084b413-8c86-4cc0-89ce-0dbfc4049e9b" (UID: "2084b413-8c86-4cc0-89ce-0dbfc4049e9b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:06 crc kubenswrapper[4775]: I0321 05:06:06.816176 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:06:06 crc kubenswrapper[4775]: I0321 05:06:06.816370 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2084b413-8c86-4cc0-89ce-0dbfc4049e9b-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:06 crc kubenswrapper[4775]: I0321 05:06:06.816380 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2084b413-8c86-4cc0-89ce-0dbfc4049e9b-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:06 crc kubenswrapper[4775]: I0321 05:06:06.816392 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f77fw\" (UniqueName: \"kubernetes.io/projected/2084b413-8c86-4cc0-89ce-0dbfc4049e9b-kube-api-access-f77fw\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:06 crc kubenswrapper[4775]: E0321 05:06:06.816554 4775 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 21 05:06:06 crc kubenswrapper[4775]: E0321 05:06:06.816588 4775 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 21 05:06:06 crc kubenswrapper[4775]: E0321 05:06:06.816639 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift podName:8e93b938-c138-4cfc-a227-e1cd648ad59a nodeName:}" failed. No retries permitted until 2026-03-21 05:06:22.816621463 +0000 UTC m=+1135.793085087 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift") pod "swift-storage-0" (UID: "8e93b938-c138-4cfc-a227-e1cd648ad59a") : configmap "swift-ring-files" not found Mar 21 05:06:06 crc kubenswrapper[4775]: I0321 05:06:06.867081 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-tq6zw" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.038679 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rn42\" (UniqueName: \"kubernetes.io/projected/af1531c7-e7a2-45cd-a5b9-9678c4808315-kube-api-access-4rn42\") pod \"af1531c7-e7a2-45cd-a5b9-9678c4808315\" (UID: \"af1531c7-e7a2-45cd-a5b9-9678c4808315\") " Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.038759 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af1531c7-e7a2-45cd-a5b9-9678c4808315-operator-scripts\") pod \"af1531c7-e7a2-45cd-a5b9-9678c4808315\" (UID: \"af1531c7-e7a2-45cd-a5b9-9678c4808315\") " Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.039616 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af1531c7-e7a2-45cd-a5b9-9678c4808315-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "af1531c7-e7a2-45cd-a5b9-9678c4808315" (UID: "af1531c7-e7a2-45cd-a5b9-9678c4808315"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.042716 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af1531c7-e7a2-45cd-a5b9-9678c4808315-kube-api-access-4rn42" (OuterVolumeSpecName: "kube-api-access-4rn42") pod "af1531c7-e7a2-45cd-a5b9-9678c4808315" (UID: "af1531c7-e7a2-45cd-a5b9-9678c4808315"). InnerVolumeSpecName "kube-api-access-4rn42". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.141039 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rn42\" (UniqueName: \"kubernetes.io/projected/af1531c7-e7a2-45cd-a5b9-9678c4808315-kube-api-access-4rn42\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.142250 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af1531c7-e7a2-45cd-a5b9-9678c4808315-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.338413 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-fmfms"] Mar 21 05:06:07 crc kubenswrapper[4775]: E0321 05:06:07.338780 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2084b413-8c86-4cc0-89ce-0dbfc4049e9b" containerName="dnsmasq-dns" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.338801 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2084b413-8c86-4cc0-89ce-0dbfc4049e9b" containerName="dnsmasq-dns" Mar 21 05:06:07 crc kubenswrapper[4775]: E0321 05:06:07.338827 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af1531c7-e7a2-45cd-a5b9-9678c4808315" containerName="mariadb-account-create-update" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.338835 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="af1531c7-e7a2-45cd-a5b9-9678c4808315" containerName="mariadb-account-create-update" Mar 21 05:06:07 crc kubenswrapper[4775]: E0321 05:06:07.338852 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2084b413-8c86-4cc0-89ce-0dbfc4049e9b" containerName="init" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.338859 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2084b413-8c86-4cc0-89ce-0dbfc4049e9b" containerName="init" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.339001 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2084b413-8c86-4cc0-89ce-0dbfc4049e9b" containerName="dnsmasq-dns" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.339015 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="af1531c7-e7a2-45cd-a5b9-9678c4808315" containerName="mariadb-account-create-update" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.339530 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-fmfms" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.353200 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-fmfms"] Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.446680 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a151b65-5187-48d2-a29a-30eae42c179c-operator-scripts\") pod \"glance-db-create-fmfms\" (UID: \"3a151b65-5187-48d2-a29a-30eae42c179c\") " pod="openstack/glance-db-create-fmfms" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.446731 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8kbf\" (UniqueName: \"kubernetes.io/projected/3a151b65-5187-48d2-a29a-30eae42c179c-kube-api-access-w8kbf\") pod \"glance-db-create-fmfms\" (UID: \"3a151b65-5187-48d2-a29a-30eae42c179c\") " pod="openstack/glance-db-create-fmfms" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.448819 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-6de1-account-create-update-bkzxl"] Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.450277 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6de1-account-create-update-bkzxl" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.454025 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.457848 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6de1-account-create-update-bkzxl"] Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.548899 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a151b65-5187-48d2-a29a-30eae42c179c-operator-scripts\") pod \"glance-db-create-fmfms\" (UID: \"3a151b65-5187-48d2-a29a-30eae42c179c\") " pod="openstack/glance-db-create-fmfms" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.548946 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8c83c73-8982-4c8e-bdeb-204a4874162c-operator-scripts\") pod \"glance-6de1-account-create-update-bkzxl\" (UID: \"a8c83c73-8982-4c8e-bdeb-204a4874162c\") " pod="openstack/glance-6de1-account-create-update-bkzxl" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.548985 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5pph\" (UniqueName: \"kubernetes.io/projected/a8c83c73-8982-4c8e-bdeb-204a4874162c-kube-api-access-b5pph\") pod \"glance-6de1-account-create-update-bkzxl\" (UID: \"a8c83c73-8982-4c8e-bdeb-204a4874162c\") " pod="openstack/glance-6de1-account-create-update-bkzxl" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.549020 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8kbf\" (UniqueName: \"kubernetes.io/projected/3a151b65-5187-48d2-a29a-30eae42c179c-kube-api-access-w8kbf\") pod \"glance-db-create-fmfms\" (UID: \"3a151b65-5187-48d2-a29a-30eae42c179c\") " pod="openstack/glance-db-create-fmfms" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.549858 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a151b65-5187-48d2-a29a-30eae42c179c-operator-scripts\") pod \"glance-db-create-fmfms\" (UID: \"3a151b65-5187-48d2-a29a-30eae42c179c\") " pod="openstack/glance-db-create-fmfms" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.564789 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8kbf\" (UniqueName: \"kubernetes.io/projected/3a151b65-5187-48d2-a29a-30eae42c179c-kube-api-access-w8kbf\") pod \"glance-db-create-fmfms\" (UID: \"3a151b65-5187-48d2-a29a-30eae42c179c\") " pod="openstack/glance-db-create-fmfms" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.612244 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-tq6zw" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.612244 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-tq6zw" event={"ID":"af1531c7-e7a2-45cd-a5b9-9678c4808315","Type":"ContainerDied","Data":"490b9754f17abc231dc24b2f80cf9533e36998bd95e31c2f91871cb4975c2c6c"} Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.612255 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-mn5xb" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.612322 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="490b9754f17abc231dc24b2f80cf9533e36998bd95e31c2f91871cb4975c2c6c" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.650899 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8c83c73-8982-4c8e-bdeb-204a4874162c-operator-scripts\") pod \"glance-6de1-account-create-update-bkzxl\" (UID: \"a8c83c73-8982-4c8e-bdeb-204a4874162c\") " pod="openstack/glance-6de1-account-create-update-bkzxl" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.650945 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5pph\" (UniqueName: \"kubernetes.io/projected/a8c83c73-8982-4c8e-bdeb-204a4874162c-kube-api-access-b5pph\") pod \"glance-6de1-account-create-update-bkzxl\" (UID: \"a8c83c73-8982-4c8e-bdeb-204a4874162c\") " pod="openstack/glance-6de1-account-create-update-bkzxl" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.651970 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8c83c73-8982-4c8e-bdeb-204a4874162c-operator-scripts\") pod \"glance-6de1-account-create-update-bkzxl\" (UID: \"a8c83c73-8982-4c8e-bdeb-204a4874162c\") " pod="openstack/glance-6de1-account-create-update-bkzxl" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.652402 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mn5xb"] Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.656402 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-fmfms" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.658539 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mn5xb"] Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.673826 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5pph\" (UniqueName: \"kubernetes.io/projected/a8c83c73-8982-4c8e-bdeb-204a4874162c-kube-api-access-b5pph\") pod \"glance-6de1-account-create-update-bkzxl\" (UID: \"a8c83c73-8982-4c8e-bdeb-204a4874162c\") " pod="openstack/glance-6de1-account-create-update-bkzxl" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.675288 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2084b413-8c86-4cc0-89ce-0dbfc4049e9b" path="/var/lib/kubelet/pods/2084b413-8c86-4cc0-89ce-0dbfc4049e9b/volumes" Mar 21 05:06:07 crc kubenswrapper[4775]: I0321 05:06:07.763651 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6de1-account-create-update-bkzxl" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.130748 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-fmfms"] Mar 21 05:06:08 crc kubenswrapper[4775]: W0321 05:06:08.132197 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a151b65_5187_48d2_a29a_30eae42c179c.slice/crio-eed09522f9eaa338fb2dacfffe92955dd1a12162b8fc1fd1c6361c9bf7b0db23 WatchSource:0}: Error finding container eed09522f9eaa338fb2dacfffe92955dd1a12162b8fc1fd1c6361c9bf7b0db23: Status 404 returned error can't find the container with id eed09522f9eaa338fb2dacfffe92955dd1a12162b8fc1fd1c6361c9bf7b0db23 Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.211657 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6de1-account-create-update-bkzxl"] Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.243034 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-j56nc"] Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.244428 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-j56nc" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.256901 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-j56nc"] Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.305524 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-594f-account-create-update-9ntmw"] Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.306445 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-594f-account-create-update-9ntmw" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.309033 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.314350 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-594f-account-create-update-9ntmw"] Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.367043 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lm6t\" (UniqueName: \"kubernetes.io/projected/17e15fc9-dccb-44f6-9266-2bfe23d9e224-kube-api-access-6lm6t\") pod \"keystone-db-create-j56nc\" (UID: \"17e15fc9-dccb-44f6-9266-2bfe23d9e224\") " pod="openstack/keystone-db-create-j56nc" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.367350 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17e15fc9-dccb-44f6-9266-2bfe23d9e224-operator-scripts\") pod \"keystone-db-create-j56nc\" (UID: \"17e15fc9-dccb-44f6-9266-2bfe23d9e224\") " pod="openstack/keystone-db-create-j56nc" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.419719 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-429fj"] Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.421166 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-429fj" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.428806 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-429fj"] Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.468599 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lm6t\" (UniqueName: \"kubernetes.io/projected/17e15fc9-dccb-44f6-9266-2bfe23d9e224-kube-api-access-6lm6t\") pod \"keystone-db-create-j56nc\" (UID: \"17e15fc9-dccb-44f6-9266-2bfe23d9e224\") " pod="openstack/keystone-db-create-j56nc" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.468728 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbh4h\" (UniqueName: \"kubernetes.io/projected/88db3052-d0c3-4f00-a116-aea162a7790b-kube-api-access-tbh4h\") pod \"keystone-594f-account-create-update-9ntmw\" (UID: \"88db3052-d0c3-4f00-a116-aea162a7790b\") " pod="openstack/keystone-594f-account-create-update-9ntmw" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.468766 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17e15fc9-dccb-44f6-9266-2bfe23d9e224-operator-scripts\") pod \"keystone-db-create-j56nc\" (UID: \"17e15fc9-dccb-44f6-9266-2bfe23d9e224\") " pod="openstack/keystone-db-create-j56nc" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.468829 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88db3052-d0c3-4f00-a116-aea162a7790b-operator-scripts\") pod \"keystone-594f-account-create-update-9ntmw\" (UID: \"88db3052-d0c3-4f00-a116-aea162a7790b\") " pod="openstack/keystone-594f-account-create-update-9ntmw" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.469675 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17e15fc9-dccb-44f6-9266-2bfe23d9e224-operator-scripts\") pod \"keystone-db-create-j56nc\" (UID: \"17e15fc9-dccb-44f6-9266-2bfe23d9e224\") " pod="openstack/keystone-db-create-j56nc" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.489484 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lm6t\" (UniqueName: \"kubernetes.io/projected/17e15fc9-dccb-44f6-9266-2bfe23d9e224-kube-api-access-6lm6t\") pod \"keystone-db-create-j56nc\" (UID: \"17e15fc9-dccb-44f6-9266-2bfe23d9e224\") " pod="openstack/keystone-db-create-j56nc" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.551744 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-41dd-account-create-update-5ltn6"] Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.552713 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-41dd-account-create-update-5ltn6" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.555198 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.563688 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-41dd-account-create-update-5ltn6"] Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.569684 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92b4f\" (UniqueName: \"kubernetes.io/projected/a5ecd8b9-b671-4181-96e9-adf3e820c8c7-kube-api-access-92b4f\") pod \"placement-db-create-429fj\" (UID: \"a5ecd8b9-b671-4181-96e9-adf3e820c8c7\") " pod="openstack/placement-db-create-429fj" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.569737 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbh4h\" (UniqueName: \"kubernetes.io/projected/88db3052-d0c3-4f00-a116-aea162a7790b-kube-api-access-tbh4h\") pod \"keystone-594f-account-create-update-9ntmw\" (UID: \"88db3052-d0c3-4f00-a116-aea162a7790b\") " pod="openstack/keystone-594f-account-create-update-9ntmw" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.569882 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88db3052-d0c3-4f00-a116-aea162a7790b-operator-scripts\") pod \"keystone-594f-account-create-update-9ntmw\" (UID: \"88db3052-d0c3-4f00-a116-aea162a7790b\") " pod="openstack/keystone-594f-account-create-update-9ntmw" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.570704 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88db3052-d0c3-4f00-a116-aea162a7790b-operator-scripts\") pod \"keystone-594f-account-create-update-9ntmw\" (UID: \"88db3052-d0c3-4f00-a116-aea162a7790b\") " pod="openstack/keystone-594f-account-create-update-9ntmw" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.570800 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5ecd8b9-b671-4181-96e9-adf3e820c8c7-operator-scripts\") pod \"placement-db-create-429fj\" (UID: \"a5ecd8b9-b671-4181-96e9-adf3e820c8c7\") " pod="openstack/placement-db-create-429fj" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.585381 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbh4h\" (UniqueName: \"kubernetes.io/projected/88db3052-d0c3-4f00-a116-aea162a7790b-kube-api-access-tbh4h\") pod \"keystone-594f-account-create-update-9ntmw\" (UID: \"88db3052-d0c3-4f00-a116-aea162a7790b\") " pod="openstack/keystone-594f-account-create-update-9ntmw" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.621748 4775 generic.go:334] "Generic (PLEG): container finished" podID="9e521c27-9d67-47bc-b6ac-74fabb543d3f" containerID="449084f0366fca2d37734d9abfa93601028eb223fcd5ecd49cab997851f8fae3" exitCode=0 Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.621810 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-kzll6" event={"ID":"9e521c27-9d67-47bc-b6ac-74fabb543d3f","Type":"ContainerDied","Data":"449084f0366fca2d37734d9abfa93601028eb223fcd5ecd49cab997851f8fae3"} Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.625014 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-fmfms" event={"ID":"3a151b65-5187-48d2-a29a-30eae42c179c","Type":"ContainerStarted","Data":"f3bbbe70d0549a7acec5479ed8da7aa8abe15ef1981c7636975b0a1dd52d35cc"} Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.625074 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-fmfms" event={"ID":"3a151b65-5187-48d2-a29a-30eae42c179c","Type":"ContainerStarted","Data":"eed09522f9eaa338fb2dacfffe92955dd1a12162b8fc1fd1c6361c9bf7b0db23"} Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.626360 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6de1-account-create-update-bkzxl" event={"ID":"a8c83c73-8982-4c8e-bdeb-204a4874162c","Type":"ContainerStarted","Data":"e4255198ffff3f324a60a06814346bd92cba238dd6b295fce65cd1387f10c2d6"} Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.626386 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6de1-account-create-update-bkzxl" event={"ID":"a8c83c73-8982-4c8e-bdeb-204a4874162c","Type":"ContainerStarted","Data":"51e33874e51c1b20e7b3774c500e497978209cb064fcdfc29426e7e72f1a0361"} Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.628387 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-j56nc" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.659084 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-fmfms" podStartSLOduration=1.659063528 podStartE2EDuration="1.659063528s" podCreationTimestamp="2026-03-21 05:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:06:08.646451091 +0000 UTC m=+1121.622914715" watchObservedRunningTime="2026-03-21 05:06:08.659063528 +0000 UTC m=+1121.635527152" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.663107 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-594f-account-create-update-9ntmw" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.667587 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-6de1-account-create-update-bkzxl" podStartSLOduration=1.667574519 podStartE2EDuration="1.667574519s" podCreationTimestamp="2026-03-21 05:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:06:08.658865762 +0000 UTC m=+1121.635329406" watchObservedRunningTime="2026-03-21 05:06:08.667574519 +0000 UTC m=+1121.644038143" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.671714 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8wxz\" (UniqueName: \"kubernetes.io/projected/1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6-kube-api-access-h8wxz\") pod \"placement-41dd-account-create-update-5ltn6\" (UID: \"1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6\") " pod="openstack/placement-41dd-account-create-update-5ltn6" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.671779 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6-operator-scripts\") pod \"placement-41dd-account-create-update-5ltn6\" (UID: \"1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6\") " pod="openstack/placement-41dd-account-create-update-5ltn6" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.671851 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5ecd8b9-b671-4181-96e9-adf3e820c8c7-operator-scripts\") pod \"placement-db-create-429fj\" (UID: \"a5ecd8b9-b671-4181-96e9-adf3e820c8c7\") " pod="openstack/placement-db-create-429fj" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.671943 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92b4f\" (UniqueName: \"kubernetes.io/projected/a5ecd8b9-b671-4181-96e9-adf3e820c8c7-kube-api-access-92b4f\") pod \"placement-db-create-429fj\" (UID: \"a5ecd8b9-b671-4181-96e9-adf3e820c8c7\") " pod="openstack/placement-db-create-429fj" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.673195 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5ecd8b9-b671-4181-96e9-adf3e820c8c7-operator-scripts\") pod \"placement-db-create-429fj\" (UID: \"a5ecd8b9-b671-4181-96e9-adf3e820c8c7\") " pod="openstack/placement-db-create-429fj" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.692385 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92b4f\" (UniqueName: \"kubernetes.io/projected/a5ecd8b9-b671-4181-96e9-adf3e820c8c7-kube-api-access-92b4f\") pod \"placement-db-create-429fj\" (UID: \"a5ecd8b9-b671-4181-96e9-adf3e820c8c7\") " pod="openstack/placement-db-create-429fj" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.747185 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-429fj" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.773434 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8wxz\" (UniqueName: \"kubernetes.io/projected/1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6-kube-api-access-h8wxz\") pod \"placement-41dd-account-create-update-5ltn6\" (UID: \"1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6\") " pod="openstack/placement-41dd-account-create-update-5ltn6" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.773498 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6-operator-scripts\") pod \"placement-41dd-account-create-update-5ltn6\" (UID: \"1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6\") " pod="openstack/placement-41dd-account-create-update-5ltn6" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.775510 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6-operator-scripts\") pod \"placement-41dd-account-create-update-5ltn6\" (UID: \"1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6\") " pod="openstack/placement-41dd-account-create-update-5ltn6" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.790788 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8wxz\" (UniqueName: \"kubernetes.io/projected/1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6-kube-api-access-h8wxz\") pod \"placement-41dd-account-create-update-5ltn6\" (UID: \"1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6\") " pod="openstack/placement-41dd-account-create-update-5ltn6" Mar 21 05:06:08 crc kubenswrapper[4775]: I0321 05:06:08.921676 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-41dd-account-create-update-5ltn6" Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.100018 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-j56nc"] Mar 21 05:06:09 crc kubenswrapper[4775]: W0321 05:06:09.100302 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17e15fc9_dccb_44f6_9266_2bfe23d9e224.slice/crio-c91792969a9636707c8b5d363c9f76819e9d8b1a4df038d739db479fdd4b1ae0 WatchSource:0}: Error finding container c91792969a9636707c8b5d363c9f76819e9d8b1a4df038d739db479fdd4b1ae0: Status 404 returned error can't find the container with id c91792969a9636707c8b5d363c9f76819e9d8b1a4df038d739db479fdd4b1ae0 Mar 21 05:06:09 crc kubenswrapper[4775]: W0321 05:06:09.101731 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88db3052_d0c3_4f00_a116_aea162a7790b.slice/crio-b730e73ba2321274902ab7bcdcdddf4bc746dc375617551c9d0547ea06be886b WatchSource:0}: Error finding container b730e73ba2321274902ab7bcdcdddf4bc746dc375617551c9d0547ea06be886b: Status 404 returned error can't find the container with id b730e73ba2321274902ab7bcdcdddf4bc746dc375617551c9d0547ea06be886b Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.107735 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-594f-account-create-update-9ntmw"] Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.228009 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-429fj"] Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.351785 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-41dd-account-create-update-5ltn6"] Mar 21 05:06:09 crc kubenswrapper[4775]: W0321 05:06:09.358581 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1623bfa5_c196_4b3f_b1ad_2e895cb6e6d6.slice/crio-fc201b1ec5f6a9dfa6006cd5ad3ef839350e233ae22e75cc42fb3c042862542a WatchSource:0}: Error finding container fc201b1ec5f6a9dfa6006cd5ad3ef839350e233ae22e75cc42fb3c042862542a: Status 404 returned error can't find the container with id fc201b1ec5f6a9dfa6006cd5ad3ef839350e233ae22e75cc42fb3c042862542a Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.635209 4775 generic.go:334] "Generic (PLEG): container finished" podID="a8c83c73-8982-4c8e-bdeb-204a4874162c" containerID="e4255198ffff3f324a60a06814346bd92cba238dd6b295fce65cd1387f10c2d6" exitCode=0 Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.635467 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6de1-account-create-update-bkzxl" event={"ID":"a8c83c73-8982-4c8e-bdeb-204a4874162c","Type":"ContainerDied","Data":"e4255198ffff3f324a60a06814346bd92cba238dd6b295fce65cd1387f10c2d6"} Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.636901 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-41dd-account-create-update-5ltn6" event={"ID":"1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6","Type":"ContainerStarted","Data":"05fed25f10a4568226dd529e50e8f49b33dd0e98b0199020ea806a0cb77f54ee"} Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.637006 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-41dd-account-create-update-5ltn6" event={"ID":"1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6","Type":"ContainerStarted","Data":"fc201b1ec5f6a9dfa6006cd5ad3ef839350e233ae22e75cc42fb3c042862542a"} Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.638578 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-j56nc" event={"ID":"17e15fc9-dccb-44f6-9266-2bfe23d9e224","Type":"ContainerStarted","Data":"13179901d0030c432b15b9cdcd3d4bcc8af1fc1231684497164b45d09b84bafe"} Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.638626 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-j56nc" event={"ID":"17e15fc9-dccb-44f6-9266-2bfe23d9e224","Type":"ContainerStarted","Data":"c91792969a9636707c8b5d363c9f76819e9d8b1a4df038d739db479fdd4b1ae0"} Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.640041 4775 generic.go:334] "Generic (PLEG): container finished" podID="3a151b65-5187-48d2-a29a-30eae42c179c" containerID="f3bbbe70d0549a7acec5479ed8da7aa8abe15ef1981c7636975b0a1dd52d35cc" exitCode=0 Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.640128 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-fmfms" event={"ID":"3a151b65-5187-48d2-a29a-30eae42c179c","Type":"ContainerDied","Data":"f3bbbe70d0549a7acec5479ed8da7aa8abe15ef1981c7636975b0a1dd52d35cc"} Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.641783 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-429fj" event={"ID":"a5ecd8b9-b671-4181-96e9-adf3e820c8c7","Type":"ContainerStarted","Data":"a887c3f98f7514e80429067de51d11f1272a6c7ef6bbe82b5e6e6de5b394be30"} Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.641806 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-429fj" event={"ID":"a5ecd8b9-b671-4181-96e9-adf3e820c8c7","Type":"ContainerStarted","Data":"54d2934244670d8ac6794aed279432efac3b71d073195ae0ffbaf75a83df79af"} Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.647011 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-594f-account-create-update-9ntmw" event={"ID":"88db3052-d0c3-4f00-a116-aea162a7790b","Type":"ContainerStarted","Data":"2b7a69f7824c3af4bbbbe389045d25783f39861d479383b702bdac257cb7b2e3"} Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.647067 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-594f-account-create-update-9ntmw" event={"ID":"88db3052-d0c3-4f00-a116-aea162a7790b","Type":"ContainerStarted","Data":"b730e73ba2321274902ab7bcdcdddf4bc746dc375617551c9d0547ea06be886b"} Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.679074 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-429fj" podStartSLOduration=1.679054453 podStartE2EDuration="1.679054453s" podCreationTimestamp="2026-03-21 05:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:06:09.672164808 +0000 UTC m=+1122.648628432" watchObservedRunningTime="2026-03-21 05:06:09.679054453 +0000 UTC m=+1122.655518077" Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.692629 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-594f-account-create-update-9ntmw" podStartSLOduration=1.692609006 podStartE2EDuration="1.692609006s" podCreationTimestamp="2026-03-21 05:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:06:09.689713644 +0000 UTC m=+1122.666177268" watchObservedRunningTime="2026-03-21 05:06:09.692609006 +0000 UTC m=+1122.669072630" Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.719002 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-j56nc" podStartSLOduration=1.718981142 podStartE2EDuration="1.718981142s" podCreationTimestamp="2026-03-21 05:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:06:09.710257615 +0000 UTC m=+1122.686721259" watchObservedRunningTime="2026-03-21 05:06:09.718981142 +0000 UTC m=+1122.695444766" Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.730904 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-41dd-account-create-update-5ltn6" podStartSLOduration=1.730882509 podStartE2EDuration="1.730882509s" podCreationTimestamp="2026-03-21 05:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:06:09.72387099 +0000 UTC m=+1122.700334634" watchObservedRunningTime="2026-03-21 05:06:09.730882509 +0000 UTC m=+1122.707346133" Mar 21 05:06:09 crc kubenswrapper[4775]: I0321 05:06:09.982415 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.097593 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9e521c27-9d67-47bc-b6ac-74fabb543d3f-dispersionconf\") pod \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.097726 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9e521c27-9d67-47bc-b6ac-74fabb543d3f-ring-data-devices\") pod \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.098389 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e521c27-9d67-47bc-b6ac-74fabb543d3f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "9e521c27-9d67-47bc-b6ac-74fabb543d3f" (UID: "9e521c27-9d67-47bc-b6ac-74fabb543d3f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.097759 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9e521c27-9d67-47bc-b6ac-74fabb543d3f-etc-swift\") pod \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.098579 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e521c27-9d67-47bc-b6ac-74fabb543d3f-combined-ca-bundle\") pod \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.098676 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9e521c27-9d67-47bc-b6ac-74fabb543d3f-swiftconf\") pod \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.098746 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r55zr\" (UniqueName: \"kubernetes.io/projected/9e521c27-9d67-47bc-b6ac-74fabb543d3f-kube-api-access-r55zr\") pod \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.098874 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e521c27-9d67-47bc-b6ac-74fabb543d3f-scripts\") pod \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\" (UID: \"9e521c27-9d67-47bc-b6ac-74fabb543d3f\") " Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.099112 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e521c27-9d67-47bc-b6ac-74fabb543d3f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "9e521c27-9d67-47bc-b6ac-74fabb543d3f" (UID: "9e521c27-9d67-47bc-b6ac-74fabb543d3f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.100069 4775 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9e521c27-9d67-47bc-b6ac-74fabb543d3f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.100110 4775 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9e521c27-9d67-47bc-b6ac-74fabb543d3f-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.103758 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e521c27-9d67-47bc-b6ac-74fabb543d3f-kube-api-access-r55zr" (OuterVolumeSpecName: "kube-api-access-r55zr") pod "9e521c27-9d67-47bc-b6ac-74fabb543d3f" (UID: "9e521c27-9d67-47bc-b6ac-74fabb543d3f"). InnerVolumeSpecName "kube-api-access-r55zr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.106882 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e521c27-9d67-47bc-b6ac-74fabb543d3f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "9e521c27-9d67-47bc-b6ac-74fabb543d3f" (UID: "9e521c27-9d67-47bc-b6ac-74fabb543d3f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.120543 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e521c27-9d67-47bc-b6ac-74fabb543d3f-scripts" (OuterVolumeSpecName: "scripts") pod "9e521c27-9d67-47bc-b6ac-74fabb543d3f" (UID: "9e521c27-9d67-47bc-b6ac-74fabb543d3f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.122299 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e521c27-9d67-47bc-b6ac-74fabb543d3f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "9e521c27-9d67-47bc-b6ac-74fabb543d3f" (UID: "9e521c27-9d67-47bc-b6ac-74fabb543d3f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.131497 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e521c27-9d67-47bc-b6ac-74fabb543d3f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e521c27-9d67-47bc-b6ac-74fabb543d3f" (UID: "9e521c27-9d67-47bc-b6ac-74fabb543d3f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.201670 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e521c27-9d67-47bc-b6ac-74fabb543d3f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.201721 4775 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9e521c27-9d67-47bc-b6ac-74fabb543d3f-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.201737 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r55zr\" (UniqueName: \"kubernetes.io/projected/9e521c27-9d67-47bc-b6ac-74fabb543d3f-kube-api-access-r55zr\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.201754 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e521c27-9d67-47bc-b6ac-74fabb543d3f-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.201767 4775 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9e521c27-9d67-47bc-b6ac-74fabb543d3f-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.673790 4775 generic.go:334] "Generic (PLEG): container finished" podID="17e15fc9-dccb-44f6-9266-2bfe23d9e224" containerID="13179901d0030c432b15b9cdcd3d4bcc8af1fc1231684497164b45d09b84bafe" exitCode=0 Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.673895 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-j56nc" event={"ID":"17e15fc9-dccb-44f6-9266-2bfe23d9e224","Type":"ContainerDied","Data":"13179901d0030c432b15b9cdcd3d4bcc8af1fc1231684497164b45d09b84bafe"} Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.682044 4775 generic.go:334] "Generic (PLEG): container finished" podID="1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6" containerID="05fed25f10a4568226dd529e50e8f49b33dd0e98b0199020ea806a0cb77f54ee" exitCode=0 Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.682155 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-41dd-account-create-update-5ltn6" event={"ID":"1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6","Type":"ContainerDied","Data":"05fed25f10a4568226dd529e50e8f49b33dd0e98b0199020ea806a0cb77f54ee"} Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.686572 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-kzll6" event={"ID":"9e521c27-9d67-47bc-b6ac-74fabb543d3f","Type":"ContainerDied","Data":"eb20ad310e635a1adb6bbb9a38d57d53e3b88c4783ef5987e437714596390e99"} Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.686620 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb20ad310e635a1adb6bbb9a38d57d53e3b88c4783ef5987e437714596390e99" Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.686677 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-kzll6" Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.692314 4775 generic.go:334] "Generic (PLEG): container finished" podID="a5ecd8b9-b671-4181-96e9-adf3e820c8c7" containerID="a887c3f98f7514e80429067de51d11f1272a6c7ef6bbe82b5e6e6de5b394be30" exitCode=0 Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.692383 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-429fj" event={"ID":"a5ecd8b9-b671-4181-96e9-adf3e820c8c7","Type":"ContainerDied","Data":"a887c3f98f7514e80429067de51d11f1272a6c7ef6bbe82b5e6e6de5b394be30"} Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.694781 4775 generic.go:334] "Generic (PLEG): container finished" podID="88db3052-d0c3-4f00-a116-aea162a7790b" containerID="2b7a69f7824c3af4bbbbe389045d25783f39861d479383b702bdac257cb7b2e3" exitCode=0 Mar 21 05:06:10 crc kubenswrapper[4775]: I0321 05:06:10.694965 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-594f-account-create-update-9ntmw" event={"ID":"88db3052-d0c3-4f00-a116-aea162a7790b","Type":"ContainerDied","Data":"2b7a69f7824c3af4bbbbe389045d25783f39861d479383b702bdac257cb7b2e3"} Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.142089 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-tq6zw"] Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.148049 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-tq6zw"] Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.190181 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6de1-account-create-update-bkzxl" Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.196387 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-fmfms" Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.324342 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a151b65-5187-48d2-a29a-30eae42c179c-operator-scripts\") pod \"3a151b65-5187-48d2-a29a-30eae42c179c\" (UID: \"3a151b65-5187-48d2-a29a-30eae42c179c\") " Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.324445 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8c83c73-8982-4c8e-bdeb-204a4874162c-operator-scripts\") pod \"a8c83c73-8982-4c8e-bdeb-204a4874162c\" (UID: \"a8c83c73-8982-4c8e-bdeb-204a4874162c\") " Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.324506 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8kbf\" (UniqueName: \"kubernetes.io/projected/3a151b65-5187-48d2-a29a-30eae42c179c-kube-api-access-w8kbf\") pod \"3a151b65-5187-48d2-a29a-30eae42c179c\" (UID: \"3a151b65-5187-48d2-a29a-30eae42c179c\") " Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.324554 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5pph\" (UniqueName: \"kubernetes.io/projected/a8c83c73-8982-4c8e-bdeb-204a4874162c-kube-api-access-b5pph\") pod \"a8c83c73-8982-4c8e-bdeb-204a4874162c\" (UID: \"a8c83c73-8982-4c8e-bdeb-204a4874162c\") " Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.325295 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8c83c73-8982-4c8e-bdeb-204a4874162c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a8c83c73-8982-4c8e-bdeb-204a4874162c" (UID: "a8c83c73-8982-4c8e-bdeb-204a4874162c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.325587 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8c83c73-8982-4c8e-bdeb-204a4874162c-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.325763 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a151b65-5187-48d2-a29a-30eae42c179c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3a151b65-5187-48d2-a29a-30eae42c179c" (UID: "3a151b65-5187-48d2-a29a-30eae42c179c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.330523 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8c83c73-8982-4c8e-bdeb-204a4874162c-kube-api-access-b5pph" (OuterVolumeSpecName: "kube-api-access-b5pph") pod "a8c83c73-8982-4c8e-bdeb-204a4874162c" (UID: "a8c83c73-8982-4c8e-bdeb-204a4874162c"). InnerVolumeSpecName "kube-api-access-b5pph". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.330867 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a151b65-5187-48d2-a29a-30eae42c179c-kube-api-access-w8kbf" (OuterVolumeSpecName: "kube-api-access-w8kbf") pod "3a151b65-5187-48d2-a29a-30eae42c179c" (UID: "3a151b65-5187-48d2-a29a-30eae42c179c"). InnerVolumeSpecName "kube-api-access-w8kbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.426635 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a151b65-5187-48d2-a29a-30eae42c179c-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.426666 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8kbf\" (UniqueName: \"kubernetes.io/projected/3a151b65-5187-48d2-a29a-30eae42c179c-kube-api-access-w8kbf\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.426676 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5pph\" (UniqueName: \"kubernetes.io/projected/a8c83c73-8982-4c8e-bdeb-204a4874162c-kube-api-access-b5pph\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.677899 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af1531c7-e7a2-45cd-a5b9-9678c4808315" path="/var/lib/kubelet/pods/af1531c7-e7a2-45cd-a5b9-9678c4808315/volumes" Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.710822 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6de1-account-create-update-bkzxl" Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.710842 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6de1-account-create-update-bkzxl" event={"ID":"a8c83c73-8982-4c8e-bdeb-204a4874162c","Type":"ContainerDied","Data":"51e33874e51c1b20e7b3774c500e497978209cb064fcdfc29426e7e72f1a0361"} Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.711266 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51e33874e51c1b20e7b3774c500e497978209cb064fcdfc29426e7e72f1a0361" Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.713575 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-fmfms" Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.713591 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-fmfms" event={"ID":"3a151b65-5187-48d2-a29a-30eae42c179c","Type":"ContainerDied","Data":"eed09522f9eaa338fb2dacfffe92955dd1a12162b8fc1fd1c6361c9bf7b0db23"} Mar 21 05:06:11 crc kubenswrapper[4775]: I0321 05:06:11.713658 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eed09522f9eaa338fb2dacfffe92955dd1a12162b8fc1fd1c6361c9bf7b0db23" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.002743 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-429fj" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.069904 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92b4f\" (UniqueName: \"kubernetes.io/projected/a5ecd8b9-b671-4181-96e9-adf3e820c8c7-kube-api-access-92b4f\") pod \"a5ecd8b9-b671-4181-96e9-adf3e820c8c7\" (UID: \"a5ecd8b9-b671-4181-96e9-adf3e820c8c7\") " Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.070038 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5ecd8b9-b671-4181-96e9-adf3e820c8c7-operator-scripts\") pod \"a5ecd8b9-b671-4181-96e9-adf3e820c8c7\" (UID: \"a5ecd8b9-b671-4181-96e9-adf3e820c8c7\") " Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.071357 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5ecd8b9-b671-4181-96e9-adf3e820c8c7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a5ecd8b9-b671-4181-96e9-adf3e820c8c7" (UID: "a5ecd8b9-b671-4181-96e9-adf3e820c8c7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.105525 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5ecd8b9-b671-4181-96e9-adf3e820c8c7-kube-api-access-92b4f" (OuterVolumeSpecName: "kube-api-access-92b4f") pod "a5ecd8b9-b671-4181-96e9-adf3e820c8c7" (UID: "a5ecd8b9-b671-4181-96e9-adf3e820c8c7"). InnerVolumeSpecName "kube-api-access-92b4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.171322 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92b4f\" (UniqueName: \"kubernetes.io/projected/a5ecd8b9-b671-4181-96e9-adf3e820c8c7-kube-api-access-92b4f\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.171353 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5ecd8b9-b671-4181-96e9-adf3e820c8c7-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.265181 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-j56nc" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.268601 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-41dd-account-create-update-5ltn6" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.272221 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lm6t\" (UniqueName: \"kubernetes.io/projected/17e15fc9-dccb-44f6-9266-2bfe23d9e224-kube-api-access-6lm6t\") pod \"17e15fc9-dccb-44f6-9266-2bfe23d9e224\" (UID: \"17e15fc9-dccb-44f6-9266-2bfe23d9e224\") " Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.272292 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6-operator-scripts\") pod \"1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6\" (UID: \"1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6\") " Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.272311 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17e15fc9-dccb-44f6-9266-2bfe23d9e224-operator-scripts\") pod \"17e15fc9-dccb-44f6-9266-2bfe23d9e224\" (UID: \"17e15fc9-dccb-44f6-9266-2bfe23d9e224\") " Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.272329 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8wxz\" (UniqueName: \"kubernetes.io/projected/1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6-kube-api-access-h8wxz\") pod \"1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6\" (UID: \"1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6\") " Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.273517 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6" (UID: "1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.273579 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17e15fc9-dccb-44f6-9266-2bfe23d9e224-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "17e15fc9-dccb-44f6-9266-2bfe23d9e224" (UID: "17e15fc9-dccb-44f6-9266-2bfe23d9e224"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.275879 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6-kube-api-access-h8wxz" (OuterVolumeSpecName: "kube-api-access-h8wxz") pod "1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6" (UID: "1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6"). InnerVolumeSpecName "kube-api-access-h8wxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.279592 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17e15fc9-dccb-44f6-9266-2bfe23d9e224-kube-api-access-6lm6t" (OuterVolumeSpecName: "kube-api-access-6lm6t") pod "17e15fc9-dccb-44f6-9266-2bfe23d9e224" (UID: "17e15fc9-dccb-44f6-9266-2bfe23d9e224"). InnerVolumeSpecName "kube-api-access-6lm6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.373454 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lm6t\" (UniqueName: \"kubernetes.io/projected/17e15fc9-dccb-44f6-9266-2bfe23d9e224-kube-api-access-6lm6t\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.373487 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.373496 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17e15fc9-dccb-44f6-9266-2bfe23d9e224-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.373504 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8wxz\" (UniqueName: \"kubernetes.io/projected/1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6-kube-api-access-h8wxz\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.397272 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-594f-account-create-update-9ntmw" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.474453 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbh4h\" (UniqueName: \"kubernetes.io/projected/88db3052-d0c3-4f00-a116-aea162a7790b-kube-api-access-tbh4h\") pod \"88db3052-d0c3-4f00-a116-aea162a7790b\" (UID: \"88db3052-d0c3-4f00-a116-aea162a7790b\") " Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.474560 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88db3052-d0c3-4f00-a116-aea162a7790b-operator-scripts\") pod \"88db3052-d0c3-4f00-a116-aea162a7790b\" (UID: \"88db3052-d0c3-4f00-a116-aea162a7790b\") " Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.475136 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88db3052-d0c3-4f00-a116-aea162a7790b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "88db3052-d0c3-4f00-a116-aea162a7790b" (UID: "88db3052-d0c3-4f00-a116-aea162a7790b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.475787 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88db3052-d0c3-4f00-a116-aea162a7790b-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.477570 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88db3052-d0c3-4f00-a116-aea162a7790b-kube-api-access-tbh4h" (OuterVolumeSpecName: "kube-api-access-tbh4h") pod "88db3052-d0c3-4f00-a116-aea162a7790b" (UID: "88db3052-d0c3-4f00-a116-aea162a7790b"). InnerVolumeSpecName "kube-api-access-tbh4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.578433 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbh4h\" (UniqueName: \"kubernetes.io/projected/88db3052-d0c3-4f00-a116-aea162a7790b-kube-api-access-tbh4h\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.591416 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-td76j"] Mar 21 05:06:12 crc kubenswrapper[4775]: E0321 05:06:12.591763 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6" containerName="mariadb-account-create-update" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.591780 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6" containerName="mariadb-account-create-update" Mar 21 05:06:12 crc kubenswrapper[4775]: E0321 05:06:12.591802 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5ecd8b9-b671-4181-96e9-adf3e820c8c7" containerName="mariadb-database-create" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.591811 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5ecd8b9-b671-4181-96e9-adf3e820c8c7" containerName="mariadb-database-create" Mar 21 05:06:12 crc kubenswrapper[4775]: E0321 05:06:12.591822 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17e15fc9-dccb-44f6-9266-2bfe23d9e224" containerName="mariadb-database-create" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.591830 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="17e15fc9-dccb-44f6-9266-2bfe23d9e224" containerName="mariadb-database-create" Mar 21 05:06:12 crc kubenswrapper[4775]: E0321 05:06:12.591845 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a151b65-5187-48d2-a29a-30eae42c179c" containerName="mariadb-database-create" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.591853 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a151b65-5187-48d2-a29a-30eae42c179c" containerName="mariadb-database-create" Mar 21 05:06:12 crc kubenswrapper[4775]: E0321 05:06:12.591867 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8c83c73-8982-4c8e-bdeb-204a4874162c" containerName="mariadb-account-create-update" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.591874 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8c83c73-8982-4c8e-bdeb-204a4874162c" containerName="mariadb-account-create-update" Mar 21 05:06:12 crc kubenswrapper[4775]: E0321 05:06:12.591890 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e521c27-9d67-47bc-b6ac-74fabb543d3f" containerName="swift-ring-rebalance" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.591897 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e521c27-9d67-47bc-b6ac-74fabb543d3f" containerName="swift-ring-rebalance" Mar 21 05:06:12 crc kubenswrapper[4775]: E0321 05:06:12.591923 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88db3052-d0c3-4f00-a116-aea162a7790b" containerName="mariadb-account-create-update" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.591931 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="88db3052-d0c3-4f00-a116-aea162a7790b" containerName="mariadb-account-create-update" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.592131 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a151b65-5187-48d2-a29a-30eae42c179c" containerName="mariadb-database-create" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.592146 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6" containerName="mariadb-account-create-update" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.592159 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8c83c73-8982-4c8e-bdeb-204a4874162c" containerName="mariadb-account-create-update" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.592169 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="88db3052-d0c3-4f00-a116-aea162a7790b" containerName="mariadb-account-create-update" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.592182 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5ecd8b9-b671-4181-96e9-adf3e820c8c7" containerName="mariadb-database-create" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.592198 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e521c27-9d67-47bc-b6ac-74fabb543d3f" containerName="swift-ring-rebalance" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.592210 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="17e15fc9-dccb-44f6-9266-2bfe23d9e224" containerName="mariadb-database-create" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.592775 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-td76j" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.603652 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4x9cq" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.603862 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.604361 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-td76j"] Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.679986 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/716605f1-5111-4e7a-9591-18dfb5da1984-config-data\") pod \"glance-db-sync-td76j\" (UID: \"716605f1-5111-4e7a-9591-18dfb5da1984\") " pod="openstack/glance-db-sync-td76j" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.680084 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/716605f1-5111-4e7a-9591-18dfb5da1984-db-sync-config-data\") pod \"glance-db-sync-td76j\" (UID: \"716605f1-5111-4e7a-9591-18dfb5da1984\") " pod="openstack/glance-db-sync-td76j" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.680152 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/716605f1-5111-4e7a-9591-18dfb5da1984-combined-ca-bundle\") pod \"glance-db-sync-td76j\" (UID: \"716605f1-5111-4e7a-9591-18dfb5da1984\") " pod="openstack/glance-db-sync-td76j" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.680215 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmggl\" (UniqueName: \"kubernetes.io/projected/716605f1-5111-4e7a-9591-18dfb5da1984-kube-api-access-zmggl\") pod \"glance-db-sync-td76j\" (UID: \"716605f1-5111-4e7a-9591-18dfb5da1984\") " pod="openstack/glance-db-sync-td76j" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.722273 4775 generic.go:334] "Generic (PLEG): container finished" podID="839e915e-8197-48e9-8b69-56ac420a1eed" containerID="5a7046cba7a8f5bae4cf5ebd270104b81a62d747ce978d3680ad2d2e0a16d243" exitCode=0 Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.722360 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"839e915e-8197-48e9-8b69-56ac420a1eed","Type":"ContainerDied","Data":"5a7046cba7a8f5bae4cf5ebd270104b81a62d747ce978d3680ad2d2e0a16d243"} Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.733222 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-41dd-account-create-update-5ltn6" event={"ID":"1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6","Type":"ContainerDied","Data":"fc201b1ec5f6a9dfa6006cd5ad3ef839350e233ae22e75cc42fb3c042862542a"} Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.733272 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc201b1ec5f6a9dfa6006cd5ad3ef839350e233ae22e75cc42fb3c042862542a" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.733332 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-41dd-account-create-update-5ltn6" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.737227 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-j56nc" event={"ID":"17e15fc9-dccb-44f6-9266-2bfe23d9e224","Type":"ContainerDied","Data":"c91792969a9636707c8b5d363c9f76819e9d8b1a4df038d739db479fdd4b1ae0"} Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.737266 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c91792969a9636707c8b5d363c9f76819e9d8b1a4df038d739db479fdd4b1ae0" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.737341 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-j56nc" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.738998 4775 generic.go:334] "Generic (PLEG): container finished" podID="375fb8b7-b673-4fd7-ae51-5f82f33c196f" containerID="1c4292485e2c0ee4f0c83f962fffc48a32f990f8105220ac94e44a4691b7ff1f" exitCode=0 Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.739070 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"375fb8b7-b673-4fd7-ae51-5f82f33c196f","Type":"ContainerDied","Data":"1c4292485e2c0ee4f0c83f962fffc48a32f990f8105220ac94e44a4691b7ff1f"} Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.744131 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-429fj" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.744205 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-429fj" event={"ID":"a5ecd8b9-b671-4181-96e9-adf3e820c8c7","Type":"ContainerDied","Data":"54d2934244670d8ac6794aed279432efac3b71d073195ae0ffbaf75a83df79af"} Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.744251 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54d2934244670d8ac6794aed279432efac3b71d073195ae0ffbaf75a83df79af" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.747357 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-594f-account-create-update-9ntmw" event={"ID":"88db3052-d0c3-4f00-a116-aea162a7790b","Type":"ContainerDied","Data":"b730e73ba2321274902ab7bcdcdddf4bc746dc375617551c9d0547ea06be886b"} Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.747505 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b730e73ba2321274902ab7bcdcdddf4bc746dc375617551c9d0547ea06be886b" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.747393 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-594f-account-create-update-9ntmw" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.783085 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/716605f1-5111-4e7a-9591-18dfb5da1984-combined-ca-bundle\") pod \"glance-db-sync-td76j\" (UID: \"716605f1-5111-4e7a-9591-18dfb5da1984\") " pod="openstack/glance-db-sync-td76j" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.783190 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmggl\" (UniqueName: \"kubernetes.io/projected/716605f1-5111-4e7a-9591-18dfb5da1984-kube-api-access-zmggl\") pod \"glance-db-sync-td76j\" (UID: \"716605f1-5111-4e7a-9591-18dfb5da1984\") " pod="openstack/glance-db-sync-td76j" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.783247 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/716605f1-5111-4e7a-9591-18dfb5da1984-config-data\") pod \"glance-db-sync-td76j\" (UID: \"716605f1-5111-4e7a-9591-18dfb5da1984\") " pod="openstack/glance-db-sync-td76j" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.783328 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/716605f1-5111-4e7a-9591-18dfb5da1984-db-sync-config-data\") pod \"glance-db-sync-td76j\" (UID: \"716605f1-5111-4e7a-9591-18dfb5da1984\") " pod="openstack/glance-db-sync-td76j" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.791974 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/716605f1-5111-4e7a-9591-18dfb5da1984-db-sync-config-data\") pod \"glance-db-sync-td76j\" (UID: \"716605f1-5111-4e7a-9591-18dfb5da1984\") " pod="openstack/glance-db-sync-td76j" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.798919 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/716605f1-5111-4e7a-9591-18dfb5da1984-combined-ca-bundle\") pod \"glance-db-sync-td76j\" (UID: \"716605f1-5111-4e7a-9591-18dfb5da1984\") " pod="openstack/glance-db-sync-td76j" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.800842 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/716605f1-5111-4e7a-9591-18dfb5da1984-config-data\") pod \"glance-db-sync-td76j\" (UID: \"716605f1-5111-4e7a-9591-18dfb5da1984\") " pod="openstack/glance-db-sync-td76j" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.801516 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmggl\" (UniqueName: \"kubernetes.io/projected/716605f1-5111-4e7a-9591-18dfb5da1984-kube-api-access-zmggl\") pod \"glance-db-sync-td76j\" (UID: \"716605f1-5111-4e7a-9591-18dfb5da1984\") " pod="openstack/glance-db-sync-td76j" Mar 21 05:06:12 crc kubenswrapper[4775]: I0321 05:06:12.924581 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-td76j" Mar 21 05:06:13 crc kubenswrapper[4775]: I0321 05:06:13.501855 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-td76j"] Mar 21 05:06:13 crc kubenswrapper[4775]: I0321 05:06:13.757309 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"375fb8b7-b673-4fd7-ae51-5f82f33c196f","Type":"ContainerStarted","Data":"2b205fc4ed697ad5cbf0298ac1f6315a11af8ff066df4aa11b086ace7ab697d3"} Mar 21 05:06:13 crc kubenswrapper[4775]: I0321 05:06:13.757493 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:06:13 crc kubenswrapper[4775]: I0321 05:06:13.759372 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-td76j" event={"ID":"716605f1-5111-4e7a-9591-18dfb5da1984","Type":"ContainerStarted","Data":"c9c92c1914b54290237327698e3251a34ba7a8a08da38fd6bd3a1901ce249b25"} Mar 21 05:06:13 crc kubenswrapper[4775]: I0321 05:06:13.761661 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"839e915e-8197-48e9-8b69-56ac420a1eed","Type":"ContainerStarted","Data":"c1cc3f5c012ce3f528f11e083192eb0842f7dad8501df783855a594cab5449da"} Mar 21 05:06:13 crc kubenswrapper[4775]: I0321 05:06:13.761878 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 21 05:06:13 crc kubenswrapper[4775]: I0321 05:06:13.779613 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=40.423319839 podStartE2EDuration="51.779595948s" podCreationTimestamp="2026-03-21 05:05:22 +0000 UTC" firstStartedPulling="2026-03-21 05:05:27.445217766 +0000 UTC m=+1080.421681390" lastFinishedPulling="2026-03-21 05:05:38.801493875 +0000 UTC m=+1091.777957499" observedRunningTime="2026-03-21 05:06:13.778086215 +0000 UTC m=+1126.754549849" watchObservedRunningTime="2026-03-21 05:06:13.779595948 +0000 UTC m=+1126.756059572" Mar 21 05:06:13 crc kubenswrapper[4775]: I0321 05:06:13.802943 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=39.426032136 podStartE2EDuration="50.802923467s" podCreationTimestamp="2026-03-21 05:05:23 +0000 UTC" firstStartedPulling="2026-03-21 05:05:27.445726631 +0000 UTC m=+1080.422190255" lastFinishedPulling="2026-03-21 05:05:38.822617962 +0000 UTC m=+1091.799081586" observedRunningTime="2026-03-21 05:06:13.799615374 +0000 UTC m=+1126.776078998" watchObservedRunningTime="2026-03-21 05:06:13.802923467 +0000 UTC m=+1126.779387091" Mar 21 05:06:16 crc kubenswrapper[4775]: I0321 05:06:16.152545 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-drcfj"] Mar 21 05:06:16 crc kubenswrapper[4775]: I0321 05:06:16.153585 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-drcfj" Mar 21 05:06:16 crc kubenswrapper[4775]: I0321 05:06:16.155794 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 21 05:06:16 crc kubenswrapper[4775]: I0321 05:06:16.167152 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-drcfj"] Mar 21 05:06:16 crc kubenswrapper[4775]: I0321 05:06:16.235618 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4c6072b-3aa5-43ef-be24-f9b20e5095bd-operator-scripts\") pod \"root-account-create-update-drcfj\" (UID: \"b4c6072b-3aa5-43ef-be24-f9b20e5095bd\") " pod="openstack/root-account-create-update-drcfj" Mar 21 05:06:16 crc kubenswrapper[4775]: I0321 05:06:16.235704 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csdp7\" (UniqueName: \"kubernetes.io/projected/b4c6072b-3aa5-43ef-be24-f9b20e5095bd-kube-api-access-csdp7\") pod \"root-account-create-update-drcfj\" (UID: \"b4c6072b-3aa5-43ef-be24-f9b20e5095bd\") " pod="openstack/root-account-create-update-drcfj" Mar 21 05:06:16 crc kubenswrapper[4775]: I0321 05:06:16.269395 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Mar 21 05:06:16 crc kubenswrapper[4775]: I0321 05:06:16.337304 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4c6072b-3aa5-43ef-be24-f9b20e5095bd-operator-scripts\") pod \"root-account-create-update-drcfj\" (UID: \"b4c6072b-3aa5-43ef-be24-f9b20e5095bd\") " pod="openstack/root-account-create-update-drcfj" Mar 21 05:06:16 crc kubenswrapper[4775]: I0321 05:06:16.337428 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csdp7\" (UniqueName: \"kubernetes.io/projected/b4c6072b-3aa5-43ef-be24-f9b20e5095bd-kube-api-access-csdp7\") pod \"root-account-create-update-drcfj\" (UID: \"b4c6072b-3aa5-43ef-be24-f9b20e5095bd\") " pod="openstack/root-account-create-update-drcfj" Mar 21 05:06:16 crc kubenswrapper[4775]: I0321 05:06:16.338135 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4c6072b-3aa5-43ef-be24-f9b20e5095bd-operator-scripts\") pod \"root-account-create-update-drcfj\" (UID: \"b4c6072b-3aa5-43ef-be24-f9b20e5095bd\") " pod="openstack/root-account-create-update-drcfj" Mar 21 05:06:16 crc kubenswrapper[4775]: I0321 05:06:16.360369 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csdp7\" (UniqueName: \"kubernetes.io/projected/b4c6072b-3aa5-43ef-be24-f9b20e5095bd-kube-api-access-csdp7\") pod \"root-account-create-update-drcfj\" (UID: \"b4c6072b-3aa5-43ef-be24-f9b20e5095bd\") " pod="openstack/root-account-create-update-drcfj" Mar 21 05:06:16 crc kubenswrapper[4775]: I0321 05:06:16.475297 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-drcfj" Mar 21 05:06:16 crc kubenswrapper[4775]: I0321 05:06:16.981366 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-drcfj"] Mar 21 05:06:17 crc kubenswrapper[4775]: I0321 05:06:17.796380 4775 generic.go:334] "Generic (PLEG): container finished" podID="b4c6072b-3aa5-43ef-be24-f9b20e5095bd" containerID="db02c0eaade2846804c5f5491c8f4869f16bd0a0624e9878aa43eba4fd63b5ed" exitCode=0 Mar 21 05:06:17 crc kubenswrapper[4775]: I0321 05:06:17.796446 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-drcfj" event={"ID":"b4c6072b-3aa5-43ef-be24-f9b20e5095bd","Type":"ContainerDied","Data":"db02c0eaade2846804c5f5491c8f4869f16bd0a0624e9878aa43eba4fd63b5ed"} Mar 21 05:06:17 crc kubenswrapper[4775]: I0321 05:06:17.796673 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-drcfj" event={"ID":"b4c6072b-3aa5-43ef-be24-f9b20e5095bd","Type":"ContainerStarted","Data":"58910afada9e3bd245c22f7b8d6624bba6e27bd44ec98c040a3ae3b0b6c85caf"} Mar 21 05:06:17 crc kubenswrapper[4775]: I0321 05:06:17.841631 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-nmtjx" podUID="8a8e948c-2978-40c8-961b-1b010f7ea920" containerName="ovn-controller" probeResult="failure" output=< Mar 21 05:06:17 crc kubenswrapper[4775]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 21 05:06:17 crc kubenswrapper[4775]: > Mar 21 05:06:22 crc kubenswrapper[4775]: I0321 05:06:22.830164 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-nmtjx" podUID="8a8e948c-2978-40c8-961b-1b010f7ea920" containerName="ovn-controller" probeResult="failure" output=< Mar 21 05:06:22 crc kubenswrapper[4775]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 21 05:06:22 crc kubenswrapper[4775]: > Mar 21 05:06:22 crc kubenswrapper[4775]: I0321 05:06:22.876468 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:06:22 crc kubenswrapper[4775]: I0321 05:06:22.896867 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8e93b938-c138-4cfc-a227-e1cd648ad59a-etc-swift\") pod \"swift-storage-0\" (UID: \"8e93b938-c138-4cfc-a227-e1cd648ad59a\") " pod="openstack/swift-storage-0" Mar 21 05:06:22 crc kubenswrapper[4775]: I0321 05:06:22.904111 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:06:22 crc kubenswrapper[4775]: I0321 05:06:22.905967 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-frhpj" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.042143 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.131495 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-nmtjx-config-pjpqh"] Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.134564 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.137980 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.160786 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-nmtjx-config-pjpqh"] Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.288555 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrlpt\" (UniqueName: \"kubernetes.io/projected/b6526a91-d74c-4b54-9393-979ed50d1df5-kube-api-access-zrlpt\") pod \"ovn-controller-nmtjx-config-pjpqh\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.288906 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b6526a91-d74c-4b54-9393-979ed50d1df5-var-run\") pod \"ovn-controller-nmtjx-config-pjpqh\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.288939 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b6526a91-d74c-4b54-9393-979ed50d1df5-var-log-ovn\") pod \"ovn-controller-nmtjx-config-pjpqh\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.288995 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b6526a91-d74c-4b54-9393-979ed50d1df5-scripts\") pod \"ovn-controller-nmtjx-config-pjpqh\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.289057 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b6526a91-d74c-4b54-9393-979ed50d1df5-var-run-ovn\") pod \"ovn-controller-nmtjx-config-pjpqh\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.289160 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b6526a91-d74c-4b54-9393-979ed50d1df5-additional-scripts\") pod \"ovn-controller-nmtjx-config-pjpqh\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.390435 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b6526a91-d74c-4b54-9393-979ed50d1df5-additional-scripts\") pod \"ovn-controller-nmtjx-config-pjpqh\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.390503 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrlpt\" (UniqueName: \"kubernetes.io/projected/b6526a91-d74c-4b54-9393-979ed50d1df5-kube-api-access-zrlpt\") pod \"ovn-controller-nmtjx-config-pjpqh\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.390561 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b6526a91-d74c-4b54-9393-979ed50d1df5-var-run\") pod \"ovn-controller-nmtjx-config-pjpqh\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.390595 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b6526a91-d74c-4b54-9393-979ed50d1df5-var-log-ovn\") pod \"ovn-controller-nmtjx-config-pjpqh\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.390623 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b6526a91-d74c-4b54-9393-979ed50d1df5-scripts\") pod \"ovn-controller-nmtjx-config-pjpqh\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.390666 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b6526a91-d74c-4b54-9393-979ed50d1df5-var-run-ovn\") pod \"ovn-controller-nmtjx-config-pjpqh\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.390996 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b6526a91-d74c-4b54-9393-979ed50d1df5-var-run-ovn\") pod \"ovn-controller-nmtjx-config-pjpqh\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.391421 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b6526a91-d74c-4b54-9393-979ed50d1df5-var-run\") pod \"ovn-controller-nmtjx-config-pjpqh\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.391819 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b6526a91-d74c-4b54-9393-979ed50d1df5-additional-scripts\") pod \"ovn-controller-nmtjx-config-pjpqh\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.393098 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b6526a91-d74c-4b54-9393-979ed50d1df5-scripts\") pod \"ovn-controller-nmtjx-config-pjpqh\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.393213 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b6526a91-d74c-4b54-9393-979ed50d1df5-var-log-ovn\") pod \"ovn-controller-nmtjx-config-pjpqh\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.412732 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrlpt\" (UniqueName: \"kubernetes.io/projected/b6526a91-d74c-4b54-9393-979ed50d1df5-kube-api-access-zrlpt\") pod \"ovn-controller-nmtjx-config-pjpqh\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:23 crc kubenswrapper[4775]: I0321 05:06:23.472234 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:24 crc kubenswrapper[4775]: I0321 05:06:24.497497 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:06:24 crc kubenswrapper[4775]: I0321 05:06:24.564452 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.539828 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-9vtvz"] Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.541644 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9vtvz" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.550948 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-9vtvz"] Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.679916 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-rbcrx"] Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.681369 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6-operator-scripts\") pod \"cinder-db-create-9vtvz\" (UID: \"94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6\") " pod="openstack/cinder-db-create-9vtvz" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.681501 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd94l\" (UniqueName: \"kubernetes.io/projected/94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6-kube-api-access-gd94l\") pod \"cinder-db-create-9vtvz\" (UID: \"94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6\") " pod="openstack/cinder-db-create-9vtvz" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.682495 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rbcrx" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.697453 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-6ca3-account-create-update-p8m7j"] Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.698495 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6ca3-account-create-update-p8m7j" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.701849 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.706900 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rbcrx"] Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.732197 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-6ca3-account-create-update-p8m7j"] Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.772546 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-x5t7w"] Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.773818 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-x5t7w" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.781563 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-x5t7w"] Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.782725 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6-operator-scripts\") pod \"cinder-db-create-9vtvz\" (UID: \"94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6\") " pod="openstack/cinder-db-create-9vtvz" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.782789 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3e27f70-6ea1-4f2e-a42e-e65c8ba76147-operator-scripts\") pod \"cinder-6ca3-account-create-update-p8m7j\" (UID: \"d3e27f70-6ea1-4f2e-a42e-e65c8ba76147\") " pod="openstack/cinder-6ca3-account-create-update-p8m7j" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.782843 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snvvp\" (UniqueName: \"kubernetes.io/projected/32831ef2-2e09-453b-9bea-dffe7423fa37-kube-api-access-snvvp\") pod \"barbican-db-create-rbcrx\" (UID: \"32831ef2-2e09-453b-9bea-dffe7423fa37\") " pod="openstack/barbican-db-create-rbcrx" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.782887 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd94l\" (UniqueName: \"kubernetes.io/projected/94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6-kube-api-access-gd94l\") pod \"cinder-db-create-9vtvz\" (UID: \"94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6\") " pod="openstack/cinder-db-create-9vtvz" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.782931 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvw5k\" (UniqueName: \"kubernetes.io/projected/d3e27f70-6ea1-4f2e-a42e-e65c8ba76147-kube-api-access-qvw5k\") pod \"cinder-6ca3-account-create-update-p8m7j\" (UID: \"d3e27f70-6ea1-4f2e-a42e-e65c8ba76147\") " pod="openstack/cinder-6ca3-account-create-update-p8m7j" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.782999 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32831ef2-2e09-453b-9bea-dffe7423fa37-operator-scripts\") pod \"barbican-db-create-rbcrx\" (UID: \"32831ef2-2e09-453b-9bea-dffe7423fa37\") " pod="openstack/barbican-db-create-rbcrx" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.786488 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6-operator-scripts\") pod \"cinder-db-create-9vtvz\" (UID: \"94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6\") " pod="openstack/cinder-db-create-9vtvz" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.806870 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd94l\" (UniqueName: \"kubernetes.io/projected/94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6-kube-api-access-gd94l\") pod \"cinder-db-create-9vtvz\" (UID: \"94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6\") " pod="openstack/cinder-db-create-9vtvz" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.857471 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-7a73-account-create-update-jvhc8"] Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.860202 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7a73-account-create-update-jvhc8" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.862946 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.867685 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9vtvz" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.868662 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7a73-account-create-update-jvhc8"] Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.884167 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3e27f70-6ea1-4f2e-a42e-e65c8ba76147-operator-scripts\") pod \"cinder-6ca3-account-create-update-p8m7j\" (UID: \"d3e27f70-6ea1-4f2e-a42e-e65c8ba76147\") " pod="openstack/cinder-6ca3-account-create-update-p8m7j" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.884219 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snvvp\" (UniqueName: \"kubernetes.io/projected/32831ef2-2e09-453b-9bea-dffe7423fa37-kube-api-access-snvvp\") pod \"barbican-db-create-rbcrx\" (UID: \"32831ef2-2e09-453b-9bea-dffe7423fa37\") " pod="openstack/barbican-db-create-rbcrx" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.884255 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74dea94d-fe8b-49e3-9730-fbb641cafaab-operator-scripts\") pod \"neutron-db-create-x5t7w\" (UID: \"74dea94d-fe8b-49e3-9730-fbb641cafaab\") " pod="openstack/neutron-db-create-x5t7w" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.884276 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvw5k\" (UniqueName: \"kubernetes.io/projected/d3e27f70-6ea1-4f2e-a42e-e65c8ba76147-kube-api-access-qvw5k\") pod \"cinder-6ca3-account-create-update-p8m7j\" (UID: \"d3e27f70-6ea1-4f2e-a42e-e65c8ba76147\") " pod="openstack/cinder-6ca3-account-create-update-p8m7j" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.884317 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfdt5\" (UniqueName: \"kubernetes.io/projected/74dea94d-fe8b-49e3-9730-fbb641cafaab-kube-api-access-kfdt5\") pod \"neutron-db-create-x5t7w\" (UID: \"74dea94d-fe8b-49e3-9730-fbb641cafaab\") " pod="openstack/neutron-db-create-x5t7w" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.884332 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32831ef2-2e09-453b-9bea-dffe7423fa37-operator-scripts\") pod \"barbican-db-create-rbcrx\" (UID: \"32831ef2-2e09-453b-9bea-dffe7423fa37\") " pod="openstack/barbican-db-create-rbcrx" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.885469 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32831ef2-2e09-453b-9bea-dffe7423fa37-operator-scripts\") pod \"barbican-db-create-rbcrx\" (UID: \"32831ef2-2e09-453b-9bea-dffe7423fa37\") " pod="openstack/barbican-db-create-rbcrx" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.885653 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3e27f70-6ea1-4f2e-a42e-e65c8ba76147-operator-scripts\") pod \"cinder-6ca3-account-create-update-p8m7j\" (UID: \"d3e27f70-6ea1-4f2e-a42e-e65c8ba76147\") " pod="openstack/cinder-6ca3-account-create-update-p8m7j" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.919961 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snvvp\" (UniqueName: \"kubernetes.io/projected/32831ef2-2e09-453b-9bea-dffe7423fa37-kube-api-access-snvvp\") pod \"barbican-db-create-rbcrx\" (UID: \"32831ef2-2e09-453b-9bea-dffe7423fa37\") " pod="openstack/barbican-db-create-rbcrx" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.921065 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvw5k\" (UniqueName: \"kubernetes.io/projected/d3e27f70-6ea1-4f2e-a42e-e65c8ba76147-kube-api-access-qvw5k\") pod \"cinder-6ca3-account-create-update-p8m7j\" (UID: \"d3e27f70-6ea1-4f2e-a42e-e65c8ba76147\") " pod="openstack/cinder-6ca3-account-create-update-p8m7j" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.986407 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sglvc\" (UniqueName: \"kubernetes.io/projected/9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b-kube-api-access-sglvc\") pod \"barbican-7a73-account-create-update-jvhc8\" (UID: \"9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b\") " pod="openstack/barbican-7a73-account-create-update-jvhc8" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.986506 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b-operator-scripts\") pod \"barbican-7a73-account-create-update-jvhc8\" (UID: \"9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b\") " pod="openstack/barbican-7a73-account-create-update-jvhc8" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.986558 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74dea94d-fe8b-49e3-9730-fbb641cafaab-operator-scripts\") pod \"neutron-db-create-x5t7w\" (UID: \"74dea94d-fe8b-49e3-9730-fbb641cafaab\") " pod="openstack/neutron-db-create-x5t7w" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.986617 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfdt5\" (UniqueName: \"kubernetes.io/projected/74dea94d-fe8b-49e3-9730-fbb641cafaab-kube-api-access-kfdt5\") pod \"neutron-db-create-x5t7w\" (UID: \"74dea94d-fe8b-49e3-9730-fbb641cafaab\") " pod="openstack/neutron-db-create-x5t7w" Mar 21 05:06:26 crc kubenswrapper[4775]: I0321 05:06:26.987479 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74dea94d-fe8b-49e3-9730-fbb641cafaab-operator-scripts\") pod \"neutron-db-create-x5t7w\" (UID: \"74dea94d-fe8b-49e3-9730-fbb641cafaab\") " pod="openstack/neutron-db-create-x5t7w" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.006235 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rbcrx" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.014615 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfdt5\" (UniqueName: \"kubernetes.io/projected/74dea94d-fe8b-49e3-9730-fbb641cafaab-kube-api-access-kfdt5\") pod \"neutron-db-create-x5t7w\" (UID: \"74dea94d-fe8b-49e3-9730-fbb641cafaab\") " pod="openstack/neutron-db-create-x5t7w" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.020301 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6ca3-account-create-update-p8m7j" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.041308 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-dl5n2"] Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.043042 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dl5n2" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.044935 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.045195 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.045696 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-nhwlw" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.046137 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.059182 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-dl5n2"] Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.088566 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sglvc\" (UniqueName: \"kubernetes.io/projected/9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b-kube-api-access-sglvc\") pod \"barbican-7a73-account-create-update-jvhc8\" (UID: \"9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b\") " pod="openstack/barbican-7a73-account-create-update-jvhc8" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.088643 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b-operator-scripts\") pod \"barbican-7a73-account-create-update-jvhc8\" (UID: \"9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b\") " pod="openstack/barbican-7a73-account-create-update-jvhc8" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.089599 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b-operator-scripts\") pod \"barbican-7a73-account-create-update-jvhc8\" (UID: \"9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b\") " pod="openstack/barbican-7a73-account-create-update-jvhc8" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.089784 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-x5t7w" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.092241 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b34c-account-create-update-6kvs4"] Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.093628 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b34c-account-create-update-6kvs4" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.096008 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.101077 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b34c-account-create-update-6kvs4"] Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.122577 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sglvc\" (UniqueName: \"kubernetes.io/projected/9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b-kube-api-access-sglvc\") pod \"barbican-7a73-account-create-update-jvhc8\" (UID: \"9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b\") " pod="openstack/barbican-7a73-account-create-update-jvhc8" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.190204 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7a73-account-create-update-jvhc8" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.190882 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwhb4\" (UniqueName: \"kubernetes.io/projected/47698618-487f-4849-b179-34398850f0e0-kube-api-access-lwhb4\") pod \"keystone-db-sync-dl5n2\" (UID: \"47698618-487f-4849-b179-34398850f0e0\") " pod="openstack/keystone-db-sync-dl5n2" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.191256 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgwr5\" (UniqueName: \"kubernetes.io/projected/762f4dd9-9a96-4cdb-aa87-e181f5959140-kube-api-access-kgwr5\") pod \"neutron-b34c-account-create-update-6kvs4\" (UID: \"762f4dd9-9a96-4cdb-aa87-e181f5959140\") " pod="openstack/neutron-b34c-account-create-update-6kvs4" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.191378 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47698618-487f-4849-b179-34398850f0e0-config-data\") pod \"keystone-db-sync-dl5n2\" (UID: \"47698618-487f-4849-b179-34398850f0e0\") " pod="openstack/keystone-db-sync-dl5n2" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.191505 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47698618-487f-4849-b179-34398850f0e0-combined-ca-bundle\") pod \"keystone-db-sync-dl5n2\" (UID: \"47698618-487f-4849-b179-34398850f0e0\") " pod="openstack/keystone-db-sync-dl5n2" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.191620 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/762f4dd9-9a96-4cdb-aa87-e181f5959140-operator-scripts\") pod \"neutron-b34c-account-create-update-6kvs4\" (UID: \"762f4dd9-9a96-4cdb-aa87-e181f5959140\") " pod="openstack/neutron-b34c-account-create-update-6kvs4" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.292958 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwhb4\" (UniqueName: \"kubernetes.io/projected/47698618-487f-4849-b179-34398850f0e0-kube-api-access-lwhb4\") pod \"keystone-db-sync-dl5n2\" (UID: \"47698618-487f-4849-b179-34398850f0e0\") " pod="openstack/keystone-db-sync-dl5n2" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.293287 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgwr5\" (UniqueName: \"kubernetes.io/projected/762f4dd9-9a96-4cdb-aa87-e181f5959140-kube-api-access-kgwr5\") pod \"neutron-b34c-account-create-update-6kvs4\" (UID: \"762f4dd9-9a96-4cdb-aa87-e181f5959140\") " pod="openstack/neutron-b34c-account-create-update-6kvs4" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.293450 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47698618-487f-4849-b179-34398850f0e0-config-data\") pod \"keystone-db-sync-dl5n2\" (UID: \"47698618-487f-4849-b179-34398850f0e0\") " pod="openstack/keystone-db-sync-dl5n2" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.293566 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47698618-487f-4849-b179-34398850f0e0-combined-ca-bundle\") pod \"keystone-db-sync-dl5n2\" (UID: \"47698618-487f-4849-b179-34398850f0e0\") " pod="openstack/keystone-db-sync-dl5n2" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.293670 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/762f4dd9-9a96-4cdb-aa87-e181f5959140-operator-scripts\") pod \"neutron-b34c-account-create-update-6kvs4\" (UID: \"762f4dd9-9a96-4cdb-aa87-e181f5959140\") " pod="openstack/neutron-b34c-account-create-update-6kvs4" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.294561 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/762f4dd9-9a96-4cdb-aa87-e181f5959140-operator-scripts\") pod \"neutron-b34c-account-create-update-6kvs4\" (UID: \"762f4dd9-9a96-4cdb-aa87-e181f5959140\") " pod="openstack/neutron-b34c-account-create-update-6kvs4" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.297387 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47698618-487f-4849-b179-34398850f0e0-combined-ca-bundle\") pod \"keystone-db-sync-dl5n2\" (UID: \"47698618-487f-4849-b179-34398850f0e0\") " pod="openstack/keystone-db-sync-dl5n2" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.297561 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47698618-487f-4849-b179-34398850f0e0-config-data\") pod \"keystone-db-sync-dl5n2\" (UID: \"47698618-487f-4849-b179-34398850f0e0\") " pod="openstack/keystone-db-sync-dl5n2" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.311589 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgwr5\" (UniqueName: \"kubernetes.io/projected/762f4dd9-9a96-4cdb-aa87-e181f5959140-kube-api-access-kgwr5\") pod \"neutron-b34c-account-create-update-6kvs4\" (UID: \"762f4dd9-9a96-4cdb-aa87-e181f5959140\") " pod="openstack/neutron-b34c-account-create-update-6kvs4" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.315148 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwhb4\" (UniqueName: \"kubernetes.io/projected/47698618-487f-4849-b179-34398850f0e0-kube-api-access-lwhb4\") pod \"keystone-db-sync-dl5n2\" (UID: \"47698618-487f-4849-b179-34398850f0e0\") " pod="openstack/keystone-db-sync-dl5n2" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.365286 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dl5n2" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.414866 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b34c-account-create-update-6kvs4" Mar 21 05:06:27 crc kubenswrapper[4775]: I0321 05:06:27.842761 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-nmtjx" podUID="8a8e948c-2978-40c8-961b-1b010f7ea920" containerName="ovn-controller" probeResult="failure" output=< Mar 21 05:06:27 crc kubenswrapper[4775]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 21 05:06:27 crc kubenswrapper[4775]: > Mar 21 05:06:31 crc kubenswrapper[4775]: E0321 05:06:31.045790 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Mar 21 05:06:31 crc kubenswrapper[4775]: E0321 05:06:31.046683 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zmggl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-td76j_openstack(716605f1-5111-4e7a-9591-18dfb5da1984): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 05:06:31 crc kubenswrapper[4775]: E0321 05:06:31.049218 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-td76j" podUID="716605f1-5111-4e7a-9591-18dfb5da1984" Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.094716 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-drcfj" Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.256985 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csdp7\" (UniqueName: \"kubernetes.io/projected/b4c6072b-3aa5-43ef-be24-f9b20e5095bd-kube-api-access-csdp7\") pod \"b4c6072b-3aa5-43ef-be24-f9b20e5095bd\" (UID: \"b4c6072b-3aa5-43ef-be24-f9b20e5095bd\") " Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.257089 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4c6072b-3aa5-43ef-be24-f9b20e5095bd-operator-scripts\") pod \"b4c6072b-3aa5-43ef-be24-f9b20e5095bd\" (UID: \"b4c6072b-3aa5-43ef-be24-f9b20e5095bd\") " Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.258098 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4c6072b-3aa5-43ef-be24-f9b20e5095bd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b4c6072b-3aa5-43ef-be24-f9b20e5095bd" (UID: "b4c6072b-3aa5-43ef-be24-f9b20e5095bd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.258309 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4c6072b-3aa5-43ef-be24-f9b20e5095bd-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.265866 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4c6072b-3aa5-43ef-be24-f9b20e5095bd-kube-api-access-csdp7" (OuterVolumeSpecName: "kube-api-access-csdp7") pod "b4c6072b-3aa5-43ef-be24-f9b20e5095bd" (UID: "b4c6072b-3aa5-43ef-be24-f9b20e5095bd"). InnerVolumeSpecName "kube-api-access-csdp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.359723 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csdp7\" (UniqueName: \"kubernetes.io/projected/b4c6072b-3aa5-43ef-be24-f9b20e5095bd-kube-api-access-csdp7\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.700710 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-dl5n2"] Mar 21 05:06:31 crc kubenswrapper[4775]: W0321 05:06:31.703530 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47698618_487f_4849_b179_34398850f0e0.slice/crio-75aba715be05adf1dd95f2ce6be9dc16fa293f4f6be54a7933d8a4a59ff5f394 WatchSource:0}: Error finding container 75aba715be05adf1dd95f2ce6be9dc16fa293f4f6be54a7933d8a4a59ff5f394: Status 404 returned error can't find the container with id 75aba715be05adf1dd95f2ce6be9dc16fa293f4f6be54a7933d8a4a59ff5f394 Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.797522 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.869140 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-x5t7w"] Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.869208 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-6ca3-account-create-update-p8m7j"] Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.877067 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7a73-account-create-update-jvhc8"] Mar 21 05:06:31 crc kubenswrapper[4775]: W0321 05:06:31.882636 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6526a91_d74c_4b54_9393_979ed50d1df5.slice/crio-f29fab679d4709d5420d49a0724f8a93a23a7a95f4f911c1ed5c1299a2d4b274 WatchSource:0}: Error finding container f29fab679d4709d5420d49a0724f8a93a23a7a95f4f911c1ed5c1299a2d4b274: Status 404 returned error can't find the container with id f29fab679d4709d5420d49a0724f8a93a23a7a95f4f911c1ed5c1299a2d4b274 Mar 21 05:06:31 crc kubenswrapper[4775]: W0321 05:06:31.883906 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32831ef2_2e09_453b_9bea_dffe7423fa37.slice/crio-68e92a8c80718e12444d64c7c94f999e3782ac0b507338d6f918f69629f1afa8 WatchSource:0}: Error finding container 68e92a8c80718e12444d64c7c94f999e3782ac0b507338d6f918f69629f1afa8: Status 404 returned error can't find the container with id 68e92a8c80718e12444d64c7c94f999e3782ac0b507338d6f918f69629f1afa8 Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.884311 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b34c-account-create-update-6kvs4"] Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.885915 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.887243 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.893263 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-nmtjx-config-pjpqh"] Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.900316 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rbcrx"] Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.900655 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.945496 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-drcfj" event={"ID":"b4c6072b-3aa5-43ef-be24-f9b20e5095bd","Type":"ContainerDied","Data":"58910afada9e3bd245c22f7b8d6624bba6e27bd44ec98c040a3ae3b0b6c85caf"} Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.945543 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58910afada9e3bd245c22f7b8d6624bba6e27bd44ec98c040a3ae3b0b6c85caf" Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.945518 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-drcfj" Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.947561 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-x5t7w" event={"ID":"74dea94d-fe8b-49e3-9730-fbb641cafaab","Type":"ContainerStarted","Data":"b7911e665933d0bf77ca85fcb51a440cd8a4af416d859020d112bc521d195743"} Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.949695 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nmtjx-config-pjpqh" event={"ID":"b6526a91-d74c-4b54-9393-979ed50d1df5","Type":"ContainerStarted","Data":"f29fab679d4709d5420d49a0724f8a93a23a7a95f4f911c1ed5c1299a2d4b274"} Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.951095 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8e93b938-c138-4cfc-a227-e1cd648ad59a","Type":"ContainerStarted","Data":"4269ad36bf0a4e521266e7451d179b0ede5228e98c496438299414388c2d9328"} Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.952468 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b34c-account-create-update-6kvs4" event={"ID":"762f4dd9-9a96-4cdb-aa87-e181f5959140","Type":"ContainerStarted","Data":"9c1420f842e90da606dfacce213e4c6be91dcbb1dc3e5c0c7991d7ffc54d4aad"} Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.953684 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6ca3-account-create-update-p8m7j" event={"ID":"d3e27f70-6ea1-4f2e-a42e-e65c8ba76147","Type":"ContainerStarted","Data":"834109a164cebc40b9da57c2e383b99039af503bfbbdbafe954a3b8ce2a534e3"} Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.955843 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dl5n2" event={"ID":"47698618-487f-4849-b179-34398850f0e0","Type":"ContainerStarted","Data":"75aba715be05adf1dd95f2ce6be9dc16fa293f4f6be54a7933d8a4a59ff5f394"} Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.960477 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7a73-account-create-update-jvhc8" event={"ID":"9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b","Type":"ContainerStarted","Data":"43335bb44823d6794353cce5f074df6b49fb1095d0314a1a014838b6ff135074"} Mar 21 05:06:31 crc kubenswrapper[4775]: I0321 05:06:31.961962 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rbcrx" event={"ID":"32831ef2-2e09-453b-9bea-dffe7423fa37","Type":"ContainerStarted","Data":"68e92a8c80718e12444d64c7c94f999e3782ac0b507338d6f918f69629f1afa8"} Mar 21 05:06:31 crc kubenswrapper[4775]: E0321 05:06:31.963504 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-td76j" podUID="716605f1-5111-4e7a-9591-18dfb5da1984" Mar 21 05:06:32 crc kubenswrapper[4775]: I0321 05:06:32.122968 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-9vtvz"] Mar 21 05:06:32 crc kubenswrapper[4775]: W0321 05:06:32.157048 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94ac9553_1ce0_4c22_9b2a_e7f38b75a6d6.slice/crio-76ed179c77d4582dff3583fe5aa8861bdab403dbd995245cbd2608d01ae0d097 WatchSource:0}: Error finding container 76ed179c77d4582dff3583fe5aa8861bdab403dbd995245cbd2608d01ae0d097: Status 404 returned error can't find the container with id 76ed179c77d4582dff3583fe5aa8861bdab403dbd995245cbd2608d01ae0d097 Mar 21 05:06:32 crc kubenswrapper[4775]: I0321 05:06:32.828826 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-nmtjx" Mar 21 05:06:32 crc kubenswrapper[4775]: I0321 05:06:32.970845 4775 generic.go:334] "Generic (PLEG): container finished" podID="94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6" containerID="3d3602be9c806dde89f844d57f5ed730a681b0e2d28bcdfcf3fd72611d393bf0" exitCode=0 Mar 21 05:06:32 crc kubenswrapper[4775]: I0321 05:06:32.970897 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-9vtvz" event={"ID":"94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6","Type":"ContainerDied","Data":"3d3602be9c806dde89f844d57f5ed730a681b0e2d28bcdfcf3fd72611d393bf0"} Mar 21 05:06:32 crc kubenswrapper[4775]: I0321 05:06:32.970947 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-9vtvz" event={"ID":"94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6","Type":"ContainerStarted","Data":"76ed179c77d4582dff3583fe5aa8861bdab403dbd995245cbd2608d01ae0d097"} Mar 21 05:06:32 crc kubenswrapper[4775]: I0321 05:06:32.973283 4775 generic.go:334] "Generic (PLEG): container finished" podID="9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b" containerID="5d0abdc2d7bcbfcb69006c6e61f11899c9c6ba0f99f80bd6ef0c79dc4299f856" exitCode=0 Mar 21 05:06:32 crc kubenswrapper[4775]: I0321 05:06:32.973418 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7a73-account-create-update-jvhc8" event={"ID":"9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b","Type":"ContainerDied","Data":"5d0abdc2d7bcbfcb69006c6e61f11899c9c6ba0f99f80bd6ef0c79dc4299f856"} Mar 21 05:06:32 crc kubenswrapper[4775]: I0321 05:06:32.980032 4775 generic.go:334] "Generic (PLEG): container finished" podID="32831ef2-2e09-453b-9bea-dffe7423fa37" containerID="9349131633ffa81b213e3981f82d949940d821489cd39e0d255301316c6d40ec" exitCode=0 Mar 21 05:06:32 crc kubenswrapper[4775]: I0321 05:06:32.980190 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rbcrx" event={"ID":"32831ef2-2e09-453b-9bea-dffe7423fa37","Type":"ContainerDied","Data":"9349131633ffa81b213e3981f82d949940d821489cd39e0d255301316c6d40ec"} Mar 21 05:06:32 crc kubenswrapper[4775]: I0321 05:06:32.983500 4775 generic.go:334] "Generic (PLEG): container finished" podID="74dea94d-fe8b-49e3-9730-fbb641cafaab" containerID="9076af8667b17ded58ed73ceb5a5784857256d7d90cd1bdaecae68562f5cf370" exitCode=0 Mar 21 05:06:32 crc kubenswrapper[4775]: I0321 05:06:32.983611 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-x5t7w" event={"ID":"74dea94d-fe8b-49e3-9730-fbb641cafaab","Type":"ContainerDied","Data":"9076af8667b17ded58ed73ceb5a5784857256d7d90cd1bdaecae68562f5cf370"} Mar 21 05:06:32 crc kubenswrapper[4775]: I0321 05:06:32.991688 4775 generic.go:334] "Generic (PLEG): container finished" podID="b6526a91-d74c-4b54-9393-979ed50d1df5" containerID="af6715536a4cd1315ee5f22687cb60f190dfcee964ee03a3fb9b9b38cd6a20b3" exitCode=0 Mar 21 05:06:32 crc kubenswrapper[4775]: I0321 05:06:32.991721 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nmtjx-config-pjpqh" event={"ID":"b6526a91-d74c-4b54-9393-979ed50d1df5","Type":"ContainerDied","Data":"af6715536a4cd1315ee5f22687cb60f190dfcee964ee03a3fb9b9b38cd6a20b3"} Mar 21 05:06:32 crc kubenswrapper[4775]: I0321 05:06:32.993823 4775 generic.go:334] "Generic (PLEG): container finished" podID="762f4dd9-9a96-4cdb-aa87-e181f5959140" containerID="7ca2f730d3508ebef97a36404de20e26217deca533b74b90b3415ce6b5463328" exitCode=0 Mar 21 05:06:32 crc kubenswrapper[4775]: I0321 05:06:32.993891 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b34c-account-create-update-6kvs4" event={"ID":"762f4dd9-9a96-4cdb-aa87-e181f5959140","Type":"ContainerDied","Data":"7ca2f730d3508ebef97a36404de20e26217deca533b74b90b3415ce6b5463328"} Mar 21 05:06:32 crc kubenswrapper[4775]: I0321 05:06:32.995771 4775 generic.go:334] "Generic (PLEG): container finished" podID="d3e27f70-6ea1-4f2e-a42e-e65c8ba76147" containerID="612cd0e0df303a73de827afa0ae711fb6535100e6b3d945953567e8f21be3662" exitCode=0 Mar 21 05:06:32 crc kubenswrapper[4775]: I0321 05:06:32.995806 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6ca3-account-create-update-p8m7j" event={"ID":"d3e27f70-6ea1-4f2e-a42e-e65c8ba76147","Type":"ContainerDied","Data":"612cd0e0df303a73de827afa0ae711fb6535100e6b3d945953567e8f21be3662"} Mar 21 05:06:34 crc kubenswrapper[4775]: I0321 05:06:34.019068 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8e93b938-c138-4cfc-a227-e1cd648ad59a","Type":"ContainerStarted","Data":"1be1e6985bc52afd6c350d60284cefc0e2c924ef8d1e7b54694759d2674029af"} Mar 21 05:06:34 crc kubenswrapper[4775]: I0321 05:06:34.019470 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8e93b938-c138-4cfc-a227-e1cd648ad59a","Type":"ContainerStarted","Data":"25d6d9373325eab7f8bc86e0cf14ba841ba38f0d3415587346c6c57f058a5065"} Mar 21 05:06:34 crc kubenswrapper[4775]: I0321 05:06:34.019488 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8e93b938-c138-4cfc-a227-e1cd648ad59a","Type":"ContainerStarted","Data":"5838046ae2e7d79c32bd035f9302cc2264bbb95880c32b0c68b16eb1c97abe8f"} Mar 21 05:06:34 crc kubenswrapper[4775]: I0321 05:06:34.019499 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8e93b938-c138-4cfc-a227-e1cd648ad59a","Type":"ContainerStarted","Data":"45dfc348c555083ad3ba97211fa33d552ff0a2755e3d42125978389eaf750637"} Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.825226 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9vtvz" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.858693 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rbcrx" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.883307 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6ca3-account-create-update-p8m7j" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.889163 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b34c-account-create-update-6kvs4" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.890243 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgwr5\" (UniqueName: \"kubernetes.io/projected/762f4dd9-9a96-4cdb-aa87-e181f5959140-kube-api-access-kgwr5\") pod \"762f4dd9-9a96-4cdb-aa87-e181f5959140\" (UID: \"762f4dd9-9a96-4cdb-aa87-e181f5959140\") " Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.890273 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gd94l\" (UniqueName: \"kubernetes.io/projected/94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6-kube-api-access-gd94l\") pod \"94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6\" (UID: \"94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6\") " Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.890315 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6-operator-scripts\") pod \"94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6\" (UID: \"94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6\") " Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.890336 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snvvp\" (UniqueName: \"kubernetes.io/projected/32831ef2-2e09-453b-9bea-dffe7423fa37-kube-api-access-snvvp\") pod \"32831ef2-2e09-453b-9bea-dffe7423fa37\" (UID: \"32831ef2-2e09-453b-9bea-dffe7423fa37\") " Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.890351 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3e27f70-6ea1-4f2e-a42e-e65c8ba76147-operator-scripts\") pod \"d3e27f70-6ea1-4f2e-a42e-e65c8ba76147\" (UID: \"d3e27f70-6ea1-4f2e-a42e-e65c8ba76147\") " Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.890368 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/762f4dd9-9a96-4cdb-aa87-e181f5959140-operator-scripts\") pod \"762f4dd9-9a96-4cdb-aa87-e181f5959140\" (UID: \"762f4dd9-9a96-4cdb-aa87-e181f5959140\") " Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.890403 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvw5k\" (UniqueName: \"kubernetes.io/projected/d3e27f70-6ea1-4f2e-a42e-e65c8ba76147-kube-api-access-qvw5k\") pod \"d3e27f70-6ea1-4f2e-a42e-e65c8ba76147\" (UID: \"d3e27f70-6ea1-4f2e-a42e-e65c8ba76147\") " Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.890437 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32831ef2-2e09-453b-9bea-dffe7423fa37-operator-scripts\") pod \"32831ef2-2e09-453b-9bea-dffe7423fa37\" (UID: \"32831ef2-2e09-453b-9bea-dffe7423fa37\") " Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.891219 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32831ef2-2e09-453b-9bea-dffe7423fa37-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "32831ef2-2e09-453b-9bea-dffe7423fa37" (UID: "32831ef2-2e09-453b-9bea-dffe7423fa37"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.891264 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6" (UID: "94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.891758 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3e27f70-6ea1-4f2e-a42e-e65c8ba76147-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d3e27f70-6ea1-4f2e-a42e-e65c8ba76147" (UID: "d3e27f70-6ea1-4f2e-a42e-e65c8ba76147"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.891790 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/762f4dd9-9a96-4cdb-aa87-e181f5959140-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "762f4dd9-9a96-4cdb-aa87-e181f5959140" (UID: "762f4dd9-9a96-4cdb-aa87-e181f5959140"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.899325 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6-kube-api-access-gd94l" (OuterVolumeSpecName: "kube-api-access-gd94l") pod "94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6" (UID: "94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6"). InnerVolumeSpecName "kube-api-access-gd94l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.900274 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3e27f70-6ea1-4f2e-a42e-e65c8ba76147-kube-api-access-qvw5k" (OuterVolumeSpecName: "kube-api-access-qvw5k") pod "d3e27f70-6ea1-4f2e-a42e-e65c8ba76147" (UID: "d3e27f70-6ea1-4f2e-a42e-e65c8ba76147"). InnerVolumeSpecName "kube-api-access-qvw5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.900338 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/762f4dd9-9a96-4cdb-aa87-e181f5959140-kube-api-access-kgwr5" (OuterVolumeSpecName: "kube-api-access-kgwr5") pod "762f4dd9-9a96-4cdb-aa87-e181f5959140" (UID: "762f4dd9-9a96-4cdb-aa87-e181f5959140"). InnerVolumeSpecName "kube-api-access-kgwr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.904151 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32831ef2-2e09-453b-9bea-dffe7423fa37-kube-api-access-snvvp" (OuterVolumeSpecName: "kube-api-access-snvvp") pod "32831ef2-2e09-453b-9bea-dffe7423fa37" (UID: "32831ef2-2e09-453b-9bea-dffe7423fa37"). InnerVolumeSpecName "kube-api-access-snvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.904742 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-x5t7w" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.912740 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7a73-account-create-update-jvhc8" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.932655 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.991293 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sglvc\" (UniqueName: \"kubernetes.io/projected/9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b-kube-api-access-sglvc\") pod \"9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b\" (UID: \"9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b\") " Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.991353 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b6526a91-d74c-4b54-9393-979ed50d1df5-additional-scripts\") pod \"b6526a91-d74c-4b54-9393-979ed50d1df5\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.991374 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrlpt\" (UniqueName: \"kubernetes.io/projected/b6526a91-d74c-4b54-9393-979ed50d1df5-kube-api-access-zrlpt\") pod \"b6526a91-d74c-4b54-9393-979ed50d1df5\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.991393 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b6526a91-d74c-4b54-9393-979ed50d1df5-scripts\") pod \"b6526a91-d74c-4b54-9393-979ed50d1df5\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.991417 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74dea94d-fe8b-49e3-9730-fbb641cafaab-operator-scripts\") pod \"74dea94d-fe8b-49e3-9730-fbb641cafaab\" (UID: \"74dea94d-fe8b-49e3-9730-fbb641cafaab\") " Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.991443 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b6526a91-d74c-4b54-9393-979ed50d1df5-var-log-ovn\") pod \"b6526a91-d74c-4b54-9393-979ed50d1df5\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.991511 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfdt5\" (UniqueName: \"kubernetes.io/projected/74dea94d-fe8b-49e3-9730-fbb641cafaab-kube-api-access-kfdt5\") pod \"74dea94d-fe8b-49e3-9730-fbb641cafaab\" (UID: \"74dea94d-fe8b-49e3-9730-fbb641cafaab\") " Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.991531 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b-operator-scripts\") pod \"9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b\" (UID: \"9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b\") " Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.991557 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b6526a91-d74c-4b54-9393-979ed50d1df5-var-run-ovn\") pod \"b6526a91-d74c-4b54-9393-979ed50d1df5\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.991578 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b6526a91-d74c-4b54-9393-979ed50d1df5-var-run\") pod \"b6526a91-d74c-4b54-9393-979ed50d1df5\" (UID: \"b6526a91-d74c-4b54-9393-979ed50d1df5\") " Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.991845 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgwr5\" (UniqueName: \"kubernetes.io/projected/762f4dd9-9a96-4cdb-aa87-e181f5959140-kube-api-access-kgwr5\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.991867 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gd94l\" (UniqueName: \"kubernetes.io/projected/94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6-kube-api-access-gd94l\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.991878 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.991888 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snvvp\" (UniqueName: \"kubernetes.io/projected/32831ef2-2e09-453b-9bea-dffe7423fa37-kube-api-access-snvvp\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.991896 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3e27f70-6ea1-4f2e-a42e-e65c8ba76147-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.991904 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/762f4dd9-9a96-4cdb-aa87-e181f5959140-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.991913 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvw5k\" (UniqueName: \"kubernetes.io/projected/d3e27f70-6ea1-4f2e-a42e-e65c8ba76147-kube-api-access-qvw5k\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.991922 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32831ef2-2e09-453b-9bea-dffe7423fa37-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.991947 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6526a91-d74c-4b54-9393-979ed50d1df5-var-run" (OuterVolumeSpecName: "var-run") pod "b6526a91-d74c-4b54-9393-979ed50d1df5" (UID: "b6526a91-d74c-4b54-9393-979ed50d1df5"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.991976 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6526a91-d74c-4b54-9393-979ed50d1df5-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "b6526a91-d74c-4b54-9393-979ed50d1df5" (UID: "b6526a91-d74c-4b54-9393-979ed50d1df5"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.992054 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74dea94d-fe8b-49e3-9730-fbb641cafaab-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "74dea94d-fe8b-49e3-9730-fbb641cafaab" (UID: "74dea94d-fe8b-49e3-9730-fbb641cafaab"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.992215 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6526a91-d74c-4b54-9393-979ed50d1df5-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "b6526a91-d74c-4b54-9393-979ed50d1df5" (UID: "b6526a91-d74c-4b54-9393-979ed50d1df5"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.992488 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6526a91-d74c-4b54-9393-979ed50d1df5-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "b6526a91-d74c-4b54-9393-979ed50d1df5" (UID: "b6526a91-d74c-4b54-9393-979ed50d1df5"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.992567 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b" (UID: "9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.992885 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6526a91-d74c-4b54-9393-979ed50d1df5-scripts" (OuterVolumeSpecName: "scripts") pod "b6526a91-d74c-4b54-9393-979ed50d1df5" (UID: "b6526a91-d74c-4b54-9393-979ed50d1df5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.994725 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b-kube-api-access-sglvc" (OuterVolumeSpecName: "kube-api-access-sglvc") pod "9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b" (UID: "9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b"). InnerVolumeSpecName "kube-api-access-sglvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.995628 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74dea94d-fe8b-49e3-9730-fbb641cafaab-kube-api-access-kfdt5" (OuterVolumeSpecName: "kube-api-access-kfdt5") pod "74dea94d-fe8b-49e3-9730-fbb641cafaab" (UID: "74dea94d-fe8b-49e3-9730-fbb641cafaab"). InnerVolumeSpecName "kube-api-access-kfdt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:36 crc kubenswrapper[4775]: I0321 05:06:36.995667 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6526a91-d74c-4b54-9393-979ed50d1df5-kube-api-access-zrlpt" (OuterVolumeSpecName: "kube-api-access-zrlpt") pod "b6526a91-d74c-4b54-9393-979ed50d1df5" (UID: "b6526a91-d74c-4b54-9393-979ed50d1df5"). InnerVolumeSpecName "kube-api-access-zrlpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.042868 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7a73-account-create-update-jvhc8" event={"ID":"9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b","Type":"ContainerDied","Data":"43335bb44823d6794353cce5f074df6b49fb1095d0314a1a014838b6ff135074"} Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.042927 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7a73-account-create-update-jvhc8" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.042935 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43335bb44823d6794353cce5f074df6b49fb1095d0314a1a014838b6ff135074" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.044601 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rbcrx" event={"ID":"32831ef2-2e09-453b-9bea-dffe7423fa37","Type":"ContainerDied","Data":"68e92a8c80718e12444d64c7c94f999e3782ac0b507338d6f918f69629f1afa8"} Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.044635 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68e92a8c80718e12444d64c7c94f999e3782ac0b507338d6f918f69629f1afa8" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.044651 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rbcrx" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.046032 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-x5t7w" event={"ID":"74dea94d-fe8b-49e3-9730-fbb641cafaab","Type":"ContainerDied","Data":"b7911e665933d0bf77ca85fcb51a440cd8a4af416d859020d112bc521d195743"} Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.046064 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7911e665933d0bf77ca85fcb51a440cd8a4af416d859020d112bc521d195743" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.046174 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-x5t7w" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.056940 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nmtjx-config-pjpqh" event={"ID":"b6526a91-d74c-4b54-9393-979ed50d1df5","Type":"ContainerDied","Data":"f29fab679d4709d5420d49a0724f8a93a23a7a95f4f911c1ed5c1299a2d4b274"} Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.056963 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f29fab679d4709d5420d49a0724f8a93a23a7a95f4f911c1ed5c1299a2d4b274" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.057005 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nmtjx-config-pjpqh" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.060325 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b34c-account-create-update-6kvs4" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.060618 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b34c-account-create-update-6kvs4" event={"ID":"762f4dd9-9a96-4cdb-aa87-e181f5959140","Type":"ContainerDied","Data":"9c1420f842e90da606dfacce213e4c6be91dcbb1dc3e5c0c7991d7ffc54d4aad"} Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.060644 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c1420f842e90da606dfacce213e4c6be91dcbb1dc3e5c0c7991d7ffc54d4aad" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.062407 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6ca3-account-create-update-p8m7j" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.062414 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6ca3-account-create-update-p8m7j" event={"ID":"d3e27f70-6ea1-4f2e-a42e-e65c8ba76147","Type":"ContainerDied","Data":"834109a164cebc40b9da57c2e383b99039af503bfbbdbafe954a3b8ce2a534e3"} Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.062460 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="834109a164cebc40b9da57c2e383b99039af503bfbbdbafe954a3b8ce2a534e3" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.064667 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9vtvz" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.065541 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-9vtvz" event={"ID":"94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6","Type":"ContainerDied","Data":"76ed179c77d4582dff3583fe5aa8861bdab403dbd995245cbd2608d01ae0d097"} Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.065586 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76ed179c77d4582dff3583fe5aa8861bdab403dbd995245cbd2608d01ae0d097" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.068396 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dl5n2" event={"ID":"47698618-487f-4849-b179-34398850f0e0","Type":"ContainerStarted","Data":"36bd24e1b5566dbbfb969007cb73ea97d71cecb278d83fa7485e1f5061b3eb1d"} Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.090823 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-dl5n2" podStartSLOduration=6.168869552 podStartE2EDuration="11.090804816s" podCreationTimestamp="2026-03-21 05:06:26 +0000 UTC" firstStartedPulling="2026-03-21 05:06:31.705668213 +0000 UTC m=+1144.682131837" lastFinishedPulling="2026-03-21 05:06:36.627603477 +0000 UTC m=+1149.604067101" observedRunningTime="2026-03-21 05:06:37.087886354 +0000 UTC m=+1150.064349998" watchObservedRunningTime="2026-03-21 05:06:37.090804816 +0000 UTC m=+1150.067268440" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.097559 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrlpt\" (UniqueName: \"kubernetes.io/projected/b6526a91-d74c-4b54-9393-979ed50d1df5-kube-api-access-zrlpt\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.097588 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b6526a91-d74c-4b54-9393-979ed50d1df5-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.097600 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74dea94d-fe8b-49e3-9730-fbb641cafaab-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.097609 4775 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b6526a91-d74c-4b54-9393-979ed50d1df5-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.097622 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfdt5\" (UniqueName: \"kubernetes.io/projected/74dea94d-fe8b-49e3-9730-fbb641cafaab-kube-api-access-kfdt5\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.097632 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.097640 4775 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b6526a91-d74c-4b54-9393-979ed50d1df5-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.097679 4775 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b6526a91-d74c-4b54-9393-979ed50d1df5-var-run\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.097787 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sglvc\" (UniqueName: \"kubernetes.io/projected/9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b-kube-api-access-sglvc\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.097802 4775 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b6526a91-d74c-4b54-9393-979ed50d1df5-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:37 crc kubenswrapper[4775]: I0321 05:06:37.528657 4775 scope.go:117] "RemoveContainer" containerID="7d8eac46771bcd3a496da8e38bf914f7acad2a5e17c1713aea69b3a4abe8ee15" Mar 21 05:06:38 crc kubenswrapper[4775]: I0321 05:06:38.034653 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-nmtjx-config-pjpqh"] Mar 21 05:06:38 crc kubenswrapper[4775]: I0321 05:06:38.046093 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-nmtjx-config-pjpqh"] Mar 21 05:06:38 crc kubenswrapper[4775]: I0321 05:06:38.078445 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8e93b938-c138-4cfc-a227-e1cd648ad59a","Type":"ContainerStarted","Data":"990c4464a56e9fdfe73e196f6716e0777acbe6f63422d3ab79b19fdbc492224b"} Mar 21 05:06:38 crc kubenswrapper[4775]: I0321 05:06:38.078495 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8e93b938-c138-4cfc-a227-e1cd648ad59a","Type":"ContainerStarted","Data":"9193bbebc1d7ac9486ecf1ebbb5784ab20e11ffe4c1d64be6b7773fd6661b433"} Mar 21 05:06:39 crc kubenswrapper[4775]: I0321 05:06:39.093421 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8e93b938-c138-4cfc-a227-e1cd648ad59a","Type":"ContainerStarted","Data":"243042feb3954e380e0eb5b0bc34136a7c3c9b637b3f94139529cf2dc89efe25"} Mar 21 05:06:39 crc kubenswrapper[4775]: I0321 05:06:39.094088 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8e93b938-c138-4cfc-a227-e1cd648ad59a","Type":"ContainerStarted","Data":"a37a7dc2c9267a0e9e64d59e2a1136010658f04372c37deef124c10dfa468170"} Mar 21 05:06:39 crc kubenswrapper[4775]: I0321 05:06:39.672215 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6526a91-d74c-4b54-9393-979ed50d1df5" path="/var/lib/kubelet/pods/b6526a91-d74c-4b54-9393-979ed50d1df5/volumes" Mar 21 05:06:42 crc kubenswrapper[4775]: I0321 05:06:42.125188 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8e93b938-c138-4cfc-a227-e1cd648ad59a","Type":"ContainerStarted","Data":"3c5e645400033189bda6cf0d6f22c66470d7937542ebf820b4e888c1efdfd822"} Mar 21 05:06:42 crc kubenswrapper[4775]: I0321 05:06:42.125818 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8e93b938-c138-4cfc-a227-e1cd648ad59a","Type":"ContainerStarted","Data":"41f21e794ac223855a36a52e58e57a9292cbe65a75e68474ac07df421017d212"} Mar 21 05:06:42 crc kubenswrapper[4775]: I0321 05:06:42.125837 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8e93b938-c138-4cfc-a227-e1cd648ad59a","Type":"ContainerStarted","Data":"4b95d23f9850741608d8b143586b154ee4a4b9b98bdb9b7751b8aae5e94cf0f7"} Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.138002 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8e93b938-c138-4cfc-a227-e1cd648ad59a","Type":"ContainerStarted","Data":"9ccb53a7eb02a588222376fcfb6617e147754262db12f8a2ad409ec6e9924d5f"} Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.138339 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8e93b938-c138-4cfc-a227-e1cd648ad59a","Type":"ContainerStarted","Data":"576404f60529e8c45704bba2171c3da8eb1caaaac0f8a7b830c16c1f6056451e"} Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.138353 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8e93b938-c138-4cfc-a227-e1cd648ad59a","Type":"ContainerStarted","Data":"c4bbf35a341f82d4c4680a9e20ce60c9d5acbc7d6b9e3c3b86a8d877b321b119"} Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.138363 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8e93b938-c138-4cfc-a227-e1cd648ad59a","Type":"ContainerStarted","Data":"dec5ec835fe61a6d902cb8de216092907ca1f69e3b0758ed5ea70a23c42d21bc"} Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.141223 4775 generic.go:334] "Generic (PLEG): container finished" podID="47698618-487f-4849-b179-34398850f0e0" containerID="36bd24e1b5566dbbfb969007cb73ea97d71cecb278d83fa7485e1f5061b3eb1d" exitCode=0 Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.141257 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dl5n2" event={"ID":"47698618-487f-4849-b179-34398850f0e0","Type":"ContainerDied","Data":"36bd24e1b5566dbbfb969007cb73ea97d71cecb278d83fa7485e1f5061b3eb1d"} Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.176077 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=44.462804215 podStartE2EDuration="54.176051559s" podCreationTimestamp="2026-03-21 05:05:49 +0000 UTC" firstStartedPulling="2026-03-21 05:06:31.79042419 +0000 UTC m=+1144.766887824" lastFinishedPulling="2026-03-21 05:06:41.503671544 +0000 UTC m=+1154.480135168" observedRunningTime="2026-03-21 05:06:43.172756675 +0000 UTC m=+1156.149220319" watchObservedRunningTime="2026-03-21 05:06:43.176051559 +0000 UTC m=+1156.152515183" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.450174 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-jhdxv"] Mar 21 05:06:43 crc kubenswrapper[4775]: E0321 05:06:43.450773 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6526a91-d74c-4b54-9393-979ed50d1df5" containerName="ovn-config" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.450787 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6526a91-d74c-4b54-9393-979ed50d1df5" containerName="ovn-config" Mar 21 05:06:43 crc kubenswrapper[4775]: E0321 05:06:43.450801 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6" containerName="mariadb-database-create" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.450807 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6" containerName="mariadb-database-create" Mar 21 05:06:43 crc kubenswrapper[4775]: E0321 05:06:43.450814 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b" containerName="mariadb-account-create-update" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.450820 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b" containerName="mariadb-account-create-update" Mar 21 05:06:43 crc kubenswrapper[4775]: E0321 05:06:43.450836 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32831ef2-2e09-453b-9bea-dffe7423fa37" containerName="mariadb-database-create" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.450843 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="32831ef2-2e09-453b-9bea-dffe7423fa37" containerName="mariadb-database-create" Mar 21 05:06:43 crc kubenswrapper[4775]: E0321 05:06:43.450855 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="762f4dd9-9a96-4cdb-aa87-e181f5959140" containerName="mariadb-account-create-update" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.450860 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="762f4dd9-9a96-4cdb-aa87-e181f5959140" containerName="mariadb-account-create-update" Mar 21 05:06:43 crc kubenswrapper[4775]: E0321 05:06:43.450869 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3e27f70-6ea1-4f2e-a42e-e65c8ba76147" containerName="mariadb-account-create-update" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.450875 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3e27f70-6ea1-4f2e-a42e-e65c8ba76147" containerName="mariadb-account-create-update" Mar 21 05:06:43 crc kubenswrapper[4775]: E0321 05:06:43.450885 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74dea94d-fe8b-49e3-9730-fbb641cafaab" containerName="mariadb-database-create" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.450890 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="74dea94d-fe8b-49e3-9730-fbb641cafaab" containerName="mariadb-database-create" Mar 21 05:06:43 crc kubenswrapper[4775]: E0321 05:06:43.450900 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4c6072b-3aa5-43ef-be24-f9b20e5095bd" containerName="mariadb-account-create-update" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.450907 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4c6072b-3aa5-43ef-be24-f9b20e5095bd" containerName="mariadb-account-create-update" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.451040 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="32831ef2-2e09-453b-9bea-dffe7423fa37" containerName="mariadb-database-create" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.451049 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6526a91-d74c-4b54-9393-979ed50d1df5" containerName="ovn-config" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.451060 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3e27f70-6ea1-4f2e-a42e-e65c8ba76147" containerName="mariadb-account-create-update" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.451070 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6" containerName="mariadb-database-create" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.451078 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="74dea94d-fe8b-49e3-9730-fbb641cafaab" containerName="mariadb-database-create" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.451086 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b" containerName="mariadb-account-create-update" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.451096 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4c6072b-3aa5-43ef-be24-f9b20e5095bd" containerName="mariadb-account-create-update" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.451107 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="762f4dd9-9a96-4cdb-aa87-e181f5959140" containerName="mariadb-account-create-update" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.451930 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.453887 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.467871 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-jhdxv"] Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.523296 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s69ql\" (UniqueName: \"kubernetes.io/projected/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-kube-api-access-s69ql\") pod \"dnsmasq-dns-5c79d794d7-jhdxv\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.523380 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-jhdxv\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.523434 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-jhdxv\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.523461 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-jhdxv\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.523491 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-config\") pod \"dnsmasq-dns-5c79d794d7-jhdxv\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.523523 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-jhdxv\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.625570 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-jhdxv\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.625678 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-jhdxv\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.625709 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-jhdxv\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.625742 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-config\") pod \"dnsmasq-dns-5c79d794d7-jhdxv\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.625773 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-jhdxv\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.625864 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s69ql\" (UniqueName: \"kubernetes.io/projected/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-kube-api-access-s69ql\") pod \"dnsmasq-dns-5c79d794d7-jhdxv\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.628098 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-jhdxv\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.628848 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-jhdxv\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.629172 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-jhdxv\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.630136 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-jhdxv\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.630562 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-config\") pod \"dnsmasq-dns-5c79d794d7-jhdxv\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.644576 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s69ql\" (UniqueName: \"kubernetes.io/projected/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-kube-api-access-s69ql\") pod \"dnsmasq-dns-5c79d794d7-jhdxv\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:43 crc kubenswrapper[4775]: I0321 05:06:43.773605 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:44 crc kubenswrapper[4775]: I0321 05:06:44.211443 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-jhdxv"] Mar 21 05:06:44 crc kubenswrapper[4775]: W0321 05:06:44.224533 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2c4bdcc_f0d3_452d_87ac_52ce486f1c52.slice/crio-ddb13a088c51c9a2b5faad8f81230304cbd7e86c02e650cace18f83a00dc10db WatchSource:0}: Error finding container ddb13a088c51c9a2b5faad8f81230304cbd7e86c02e650cace18f83a00dc10db: Status 404 returned error can't find the container with id ddb13a088c51c9a2b5faad8f81230304cbd7e86c02e650cace18f83a00dc10db Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:44.431581 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dl5n2" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:44.542006 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47698618-487f-4849-b179-34398850f0e0-combined-ca-bundle\") pod \"47698618-487f-4849-b179-34398850f0e0\" (UID: \"47698618-487f-4849-b179-34398850f0e0\") " Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:44.542144 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47698618-487f-4849-b179-34398850f0e0-config-data\") pod \"47698618-487f-4849-b179-34398850f0e0\" (UID: \"47698618-487f-4849-b179-34398850f0e0\") " Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:44.542170 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwhb4\" (UniqueName: \"kubernetes.io/projected/47698618-487f-4849-b179-34398850f0e0-kube-api-access-lwhb4\") pod \"47698618-487f-4849-b179-34398850f0e0\" (UID: \"47698618-487f-4849-b179-34398850f0e0\") " Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:44.545931 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47698618-487f-4849-b179-34398850f0e0-kube-api-access-lwhb4" (OuterVolumeSpecName: "kube-api-access-lwhb4") pod "47698618-487f-4849-b179-34398850f0e0" (UID: "47698618-487f-4849-b179-34398850f0e0"). InnerVolumeSpecName "kube-api-access-lwhb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:44.563312 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47698618-487f-4849-b179-34398850f0e0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47698618-487f-4849-b179-34398850f0e0" (UID: "47698618-487f-4849-b179-34398850f0e0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:44.586177 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47698618-487f-4849-b179-34398850f0e0-config-data" (OuterVolumeSpecName: "config-data") pod "47698618-487f-4849-b179-34398850f0e0" (UID: "47698618-487f-4849-b179-34398850f0e0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:44.644443 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47698618-487f-4849-b179-34398850f0e0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:44.644479 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47698618-487f-4849-b179-34398850f0e0-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:44.644493 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwhb4\" (UniqueName: \"kubernetes.io/projected/47698618-487f-4849-b179-34398850f0e0-kube-api-access-lwhb4\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.158292 4775 generic.go:334] "Generic (PLEG): container finished" podID="e2c4bdcc-f0d3-452d-87ac-52ce486f1c52" containerID="11cb266c6b3aa03b4030e08a2478a4c8badfb6b932b90e369c8e339bd2194478" exitCode=0 Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.158394 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" event={"ID":"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52","Type":"ContainerDied","Data":"11cb266c6b3aa03b4030e08a2478a4c8badfb6b932b90e369c8e339bd2194478"} Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.158418 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" event={"ID":"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52","Type":"ContainerStarted","Data":"ddb13a088c51c9a2b5faad8f81230304cbd7e86c02e650cace18f83a00dc10db"} Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.160272 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dl5n2" event={"ID":"47698618-487f-4849-b179-34398850f0e0","Type":"ContainerDied","Data":"75aba715be05adf1dd95f2ce6be9dc16fa293f4f6be54a7933d8a4a59ff5f394"} Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.160302 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dl5n2" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.160304 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75aba715be05adf1dd95f2ce6be9dc16fa293f4f6be54a7933d8a4a59ff5f394" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.447763 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-jhdxv"] Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.514245 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b868669f-t7qd7"] Mar 21 05:06:45 crc kubenswrapper[4775]: E0321 05:06:45.514827 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47698618-487f-4849-b179-34398850f0e0" containerName="keystone-db-sync" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.514908 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="47698618-487f-4849-b179-34398850f0e0" containerName="keystone-db-sync" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.515234 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="47698618-487f-4849-b179-34398850f0e0" containerName="keystone-db-sync" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.516435 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.531286 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-kbtmv"] Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.532707 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.541223 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-nhwlw" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.541441 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.541326 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.541394 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.541601 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.561751 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-t7qd7"] Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.579988 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-kbtmv"] Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.668295 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-t7qd7\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.668378 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-dns-svc\") pod \"dnsmasq-dns-5b868669f-t7qd7\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.668404 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m66r\" (UniqueName: \"kubernetes.io/projected/ed132640-db0b-458b-ba3f-d5aedb284447-kube-api-access-7m66r\") pod \"dnsmasq-dns-5b868669f-t7qd7\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.668445 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-t7qd7\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.668498 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-config-data\") pod \"keystone-bootstrap-kbtmv\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.668522 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-fernet-keys\") pod \"keystone-bootstrap-kbtmv\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.668555 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znghz\" (UniqueName: \"kubernetes.io/projected/50f0ec15-decb-4879-9410-0962a21f83ae-kube-api-access-znghz\") pod \"keystone-bootstrap-kbtmv\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.668595 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-config\") pod \"dnsmasq-dns-5b868669f-t7qd7\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.668628 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-scripts\") pod \"keystone-bootstrap-kbtmv\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.668655 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-t7qd7\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.668677 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-combined-ca-bundle\") pod \"keystone-bootstrap-kbtmv\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.668706 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-credential-keys\") pod \"keystone-bootstrap-kbtmv\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.708427 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-9fd475b9c-m48br"] Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.716860 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-9fd475b9c-m48br"] Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.716982 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9fd475b9c-m48br" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.725170 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-wvjlb"] Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.726434 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wvjlb" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.735633 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.735794 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-pms25" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.738479 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.749678 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-l8qqc"] Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.752469 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.752710 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-rcxc7" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.752831 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.758334 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.760055 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.772342 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-scripts\") pod \"keystone-bootstrap-kbtmv\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.772397 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-combined-ca-bundle\") pod \"keystone-bootstrap-kbtmv\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.772424 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-t7qd7\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.772455 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-credential-keys\") pod \"keystone-bootstrap-kbtmv\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.772498 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-t7qd7\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.772554 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-dns-svc\") pod \"dnsmasq-dns-5b868669f-t7qd7\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.772576 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m66r\" (UniqueName: \"kubernetes.io/projected/ed132640-db0b-458b-ba3f-d5aedb284447-kube-api-access-7m66r\") pod \"dnsmasq-dns-5b868669f-t7qd7\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.772619 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-t7qd7\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.772670 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-config-data\") pod \"keystone-bootstrap-kbtmv\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.772694 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-fernet-keys\") pod \"keystone-bootstrap-kbtmv\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.772728 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znghz\" (UniqueName: \"kubernetes.io/projected/50f0ec15-decb-4879-9410-0962a21f83ae-kube-api-access-znghz\") pod \"keystone-bootstrap-kbtmv\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.772765 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-config\") pod \"dnsmasq-dns-5b868669f-t7qd7\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.773872 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-config\") pod \"dnsmasq-dns-5b868669f-t7qd7\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.774263 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-t7qd7\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.775334 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-t7qd7\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.775892 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-dns-svc\") pod \"dnsmasq-dns-5b868669f-t7qd7\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.810774 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-config-data\") pod \"keystone-bootstrap-kbtmv\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.811382 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-fernet-keys\") pod \"keystone-bootstrap-kbtmv\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.813246 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-t7qd7\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.851492 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-nm9jl" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.851719 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.851905 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.885334 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-combined-ca-bundle\") pod \"keystone-bootstrap-kbtmv\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.885604 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-scripts\") pod \"keystone-bootstrap-kbtmv\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.887239 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf22x\" (UniqueName: \"kubernetes.io/projected/9531f5f9-8f77-4882-b779-4210b1de81de-kube-api-access-gf22x\") pod \"neutron-db-sync-wvjlb\" (UID: \"9531f5f9-8f77-4882-b779-4210b1de81de\") " pod="openstack/neutron-db-sync-wvjlb" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.887297 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9531f5f9-8f77-4882-b779-4210b1de81de-config\") pod \"neutron-db-sync-wvjlb\" (UID: \"9531f5f9-8f77-4882-b779-4210b1de81de\") " pod="openstack/neutron-db-sync-wvjlb" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.887324 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjzgj\" (UniqueName: \"kubernetes.io/projected/f281a5f0-270c-4d95-93b1-ad265786b886-kube-api-access-jjzgj\") pod \"horizon-9fd475b9c-m48br\" (UID: \"f281a5f0-270c-4d95-93b1-ad265786b886\") " pod="openstack/horizon-9fd475b9c-m48br" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.887354 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f281a5f0-270c-4d95-93b1-ad265786b886-logs\") pod \"horizon-9fd475b9c-m48br\" (UID: \"f281a5f0-270c-4d95-93b1-ad265786b886\") " pod="openstack/horizon-9fd475b9c-m48br" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.887377 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9531f5f9-8f77-4882-b779-4210b1de81de-combined-ca-bundle\") pod \"neutron-db-sync-wvjlb\" (UID: \"9531f5f9-8f77-4882-b779-4210b1de81de\") " pod="openstack/neutron-db-sync-wvjlb" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.887416 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-scripts\") pod \"cinder-db-sync-l8qqc\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.887439 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f281a5f0-270c-4d95-93b1-ad265786b886-scripts\") pod \"horizon-9fd475b9c-m48br\" (UID: \"f281a5f0-270c-4d95-93b1-ad265786b886\") " pod="openstack/horizon-9fd475b9c-m48br" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.887489 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f281a5f0-270c-4d95-93b1-ad265786b886-config-data\") pod \"horizon-9fd475b9c-m48br\" (UID: \"f281a5f0-270c-4d95-93b1-ad265786b886\") " pod="openstack/horizon-9fd475b9c-m48br" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.887534 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-db-sync-config-data\") pod \"cinder-db-sync-l8qqc\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.887629 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-config-data\") pod \"cinder-db-sync-l8qqc\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.887924 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-764f7\" (UniqueName: \"kubernetes.io/projected/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-kube-api-access-764f7\") pod \"cinder-db-sync-l8qqc\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.888002 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f281a5f0-270c-4d95-93b1-ad265786b886-horizon-secret-key\") pod \"horizon-9fd475b9c-m48br\" (UID: \"f281a5f0-270c-4d95-93b1-ad265786b886\") " pod="openstack/horizon-9fd475b9c-m48br" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.888044 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-etc-machine-id\") pod \"cinder-db-sync-l8qqc\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.888089 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-combined-ca-bundle\") pod \"cinder-db-sync-l8qqc\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.891777 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znghz\" (UniqueName: \"kubernetes.io/projected/50f0ec15-decb-4879-9410-0962a21f83ae-kube-api-access-znghz\") pod \"keystone-bootstrap-kbtmv\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.905753 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-credential-keys\") pod \"keystone-bootstrap-kbtmv\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.929179 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-l8qqc"] Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.929720 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m66r\" (UniqueName: \"kubernetes.io/projected/ed132640-db0b-458b-ba3f-d5aedb284447-kube-api-access-7m66r\") pod \"dnsmasq-dns-5b868669f-t7qd7\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:45 crc kubenswrapper[4775]: I0321 05:06:45.953130 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-wvjlb"] Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.006932 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.006975 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f281a5f0-270c-4d95-93b1-ad265786b886-horizon-secret-key\") pod \"horizon-9fd475b9c-m48br\" (UID: \"f281a5f0-270c-4d95-93b1-ad265786b886\") " pod="openstack/horizon-9fd475b9c-m48br" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.007050 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-etc-machine-id\") pod \"cinder-db-sync-l8qqc\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.007089 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-combined-ca-bundle\") pod \"cinder-db-sync-l8qqc\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.035019 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-etc-machine-id\") pod \"cinder-db-sync-l8qqc\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.041233 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf22x\" (UniqueName: \"kubernetes.io/projected/9531f5f9-8f77-4882-b779-4210b1de81de-kube-api-access-gf22x\") pod \"neutron-db-sync-wvjlb\" (UID: \"9531f5f9-8f77-4882-b779-4210b1de81de\") " pod="openstack/neutron-db-sync-wvjlb" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.041330 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9531f5f9-8f77-4882-b779-4210b1de81de-config\") pod \"neutron-db-sync-wvjlb\" (UID: \"9531f5f9-8f77-4882-b779-4210b1de81de\") " pod="openstack/neutron-db-sync-wvjlb" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.041362 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjzgj\" (UniqueName: \"kubernetes.io/projected/f281a5f0-270c-4d95-93b1-ad265786b886-kube-api-access-jjzgj\") pod \"horizon-9fd475b9c-m48br\" (UID: \"f281a5f0-270c-4d95-93b1-ad265786b886\") " pod="openstack/horizon-9fd475b9c-m48br" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.041384 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f281a5f0-270c-4d95-93b1-ad265786b886-logs\") pod \"horizon-9fd475b9c-m48br\" (UID: \"f281a5f0-270c-4d95-93b1-ad265786b886\") " pod="openstack/horizon-9fd475b9c-m48br" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.041417 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9531f5f9-8f77-4882-b779-4210b1de81de-combined-ca-bundle\") pod \"neutron-db-sync-wvjlb\" (UID: \"9531f5f9-8f77-4882-b779-4210b1de81de\") " pod="openstack/neutron-db-sync-wvjlb" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.041473 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-scripts\") pod \"cinder-db-sync-l8qqc\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.041501 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f281a5f0-270c-4d95-93b1-ad265786b886-scripts\") pod \"horizon-9fd475b9c-m48br\" (UID: \"f281a5f0-270c-4d95-93b1-ad265786b886\") " pod="openstack/horizon-9fd475b9c-m48br" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.041568 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f281a5f0-270c-4d95-93b1-ad265786b886-config-data\") pod \"horizon-9fd475b9c-m48br\" (UID: \"f281a5f0-270c-4d95-93b1-ad265786b886\") " pod="openstack/horizon-9fd475b9c-m48br" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.041621 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-db-sync-config-data\") pod \"cinder-db-sync-l8qqc\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.041723 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-config-data\") pod \"cinder-db-sync-l8qqc\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.041866 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-764f7\" (UniqueName: \"kubernetes.io/projected/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-kube-api-access-764f7\") pod \"cinder-db-sync-l8qqc\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.044192 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.046331 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f281a5f0-270c-4d95-93b1-ad265786b886-config-data\") pod \"horizon-9fd475b9c-m48br\" (UID: \"f281a5f0-270c-4d95-93b1-ad265786b886\") " pod="openstack/horizon-9fd475b9c-m48br" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.046911 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f281a5f0-270c-4d95-93b1-ad265786b886-scripts\") pod \"horizon-9fd475b9c-m48br\" (UID: \"f281a5f0-270c-4d95-93b1-ad265786b886\") " pod="openstack/horizon-9fd475b9c-m48br" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.048777 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f281a5f0-270c-4d95-93b1-ad265786b886-horizon-secret-key\") pod \"horizon-9fd475b9c-m48br\" (UID: \"f281a5f0-270c-4d95-93b1-ad265786b886\") " pod="openstack/horizon-9fd475b9c-m48br" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.049840 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f281a5f0-270c-4d95-93b1-ad265786b886-logs\") pod \"horizon-9fd475b9c-m48br\" (UID: \"f281a5f0-270c-4d95-93b1-ad265786b886\") " pod="openstack/horizon-9fd475b9c-m48br" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.061630 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9531f5f9-8f77-4882-b779-4210b1de81de-combined-ca-bundle\") pod \"neutron-db-sync-wvjlb\" (UID: \"9531f5f9-8f77-4882-b779-4210b1de81de\") " pod="openstack/neutron-db-sync-wvjlb" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.062660 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.067832 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.072338 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-config-data\") pod \"cinder-db-sync-l8qqc\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.074413 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-scripts\") pod \"cinder-db-sync-l8qqc\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.078566 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-db-sync-config-data\") pod \"cinder-db-sync-l8qqc\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.080345 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9531f5f9-8f77-4882-b779-4210b1de81de-config\") pod \"neutron-db-sync-wvjlb\" (UID: \"9531f5f9-8f77-4882-b779-4210b1de81de\") " pod="openstack/neutron-db-sync-wvjlb" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.106352 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-combined-ca-bundle\") pod \"cinder-db-sync-l8qqc\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.106599 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf22x\" (UniqueName: \"kubernetes.io/projected/9531f5f9-8f77-4882-b779-4210b1de81de-kube-api-access-gf22x\") pod \"neutron-db-sync-wvjlb\" (UID: \"9531f5f9-8f77-4882-b779-4210b1de81de\") " pod="openstack/neutron-db-sync-wvjlb" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.108296 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-764f7\" (UniqueName: \"kubernetes.io/projected/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-kube-api-access-764f7\") pod \"cinder-db-sync-l8qqc\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.121868 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.130785 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjzgj\" (UniqueName: \"kubernetes.io/projected/f281a5f0-270c-4d95-93b1-ad265786b886-kube-api-access-jjzgj\") pod \"horizon-9fd475b9c-m48br\" (UID: \"f281a5f0-270c-4d95-93b1-ad265786b886\") " pod="openstack/horizon-9fd475b9c-m48br" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.135233 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7bc489768c-rlbc8"] Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.137022 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bc489768c-rlbc8" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.150902 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.176752 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ac5c16c-56ac-4299-ae61-a8200986ce10-run-httpd\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.177018 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ac5c16c-56ac-4299-ae61-a8200986ce10-log-httpd\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.177321 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8h8k\" (UniqueName: \"kubernetes.io/projected/8ac5c16c-56ac-4299-ae61-a8200986ce10-kube-api-access-w8h8k\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.177500 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.177594 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-scripts\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.177682 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.177779 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-config-data\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.178825 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.229833 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" event={"ID":"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52","Type":"ContainerStarted","Data":"44277cb33149a33c5c08b0bf964d1c1e6f322cfb36b55ecaba7cde47956ea895"} Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.230825 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.236596 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" podUID="e2c4bdcc-f0d3-452d-87ac-52ce486f1c52" containerName="dnsmasq-dns" containerID="cri-o://44277cb33149a33c5c08b0bf964d1c1e6f322cfb36b55ecaba7cde47956ea895" gracePeriod=10 Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.247690 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7bc489768c-rlbc8"] Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.283169 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-scripts\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.283225 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.283255 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-config-data\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.283313 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ac5c16c-56ac-4299-ae61-a8200986ce10-run-httpd\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.283336 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ac5c16c-56ac-4299-ae61-a8200986ce10-log-httpd\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.283368 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tstsp\" (UniqueName: \"kubernetes.io/projected/44b58fbc-be53-4531-9dee-05ba398f25ef-kube-api-access-tstsp\") pod \"horizon-7bc489768c-rlbc8\" (UID: \"44b58fbc-be53-4531-9dee-05ba398f25ef\") " pod="openstack/horizon-7bc489768c-rlbc8" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.283398 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/44b58fbc-be53-4531-9dee-05ba398f25ef-horizon-secret-key\") pod \"horizon-7bc489768c-rlbc8\" (UID: \"44b58fbc-be53-4531-9dee-05ba398f25ef\") " pod="openstack/horizon-7bc489768c-rlbc8" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.283420 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/44b58fbc-be53-4531-9dee-05ba398f25ef-scripts\") pod \"horizon-7bc489768c-rlbc8\" (UID: \"44b58fbc-be53-4531-9dee-05ba398f25ef\") " pod="openstack/horizon-7bc489768c-rlbc8" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.283439 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/44b58fbc-be53-4531-9dee-05ba398f25ef-config-data\") pod \"horizon-7bc489768c-rlbc8\" (UID: \"44b58fbc-be53-4531-9dee-05ba398f25ef\") " pod="openstack/horizon-7bc489768c-rlbc8" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.283464 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44b58fbc-be53-4531-9dee-05ba398f25ef-logs\") pod \"horizon-7bc489768c-rlbc8\" (UID: \"44b58fbc-be53-4531-9dee-05ba398f25ef\") " pod="openstack/horizon-7bc489768c-rlbc8" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.283503 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8h8k\" (UniqueName: \"kubernetes.io/projected/8ac5c16c-56ac-4299-ae61-a8200986ce10-kube-api-access-w8h8k\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.283552 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.284690 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ac5c16c-56ac-4299-ae61-a8200986ce10-log-httpd\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.284837 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ac5c16c-56ac-4299-ae61-a8200986ce10-run-httpd\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.285494 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-t7qd7"] Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.286042 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wvjlb" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.290026 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-scripts\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.291975 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.292035 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.293318 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-config-data\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.298972 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-dh4pj"] Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.300298 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dh4pj" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.304455 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-8q2g8" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.304599 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.304729 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.307053 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8h8k\" (UniqueName: \"kubernetes.io/projected/8ac5c16c-56ac-4299-ae61-a8200986ce10-kube-api-access-w8h8k\") pod \"ceilometer-0\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.309927 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-9cmq2"] Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.311703 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9cmq2" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.316504 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.316710 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-ph6zz" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.341331 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9cmq2"] Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.351531 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4nfhc"] Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.353094 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.364538 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-dh4pj"] Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.364957 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9fd475b9c-m48br" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.386103 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1658991f-7a2b-4ce9-a240-a940385e0b8f-db-sync-config-data\") pod \"barbican-db-sync-9cmq2\" (UID: \"1658991f-7a2b-4ce9-a240-a940385e0b8f\") " pod="openstack/barbican-db-sync-9cmq2" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.386263 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bczb\" (UniqueName: \"kubernetes.io/projected/1658991f-7a2b-4ce9-a240-a940385e0b8f-kube-api-access-9bczb\") pod \"barbican-db-sync-9cmq2\" (UID: \"1658991f-7a2b-4ce9-a240-a940385e0b8f\") " pod="openstack/barbican-db-sync-9cmq2" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.386339 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tstsp\" (UniqueName: \"kubernetes.io/projected/44b58fbc-be53-4531-9dee-05ba398f25ef-kube-api-access-tstsp\") pod \"horizon-7bc489768c-rlbc8\" (UID: \"44b58fbc-be53-4531-9dee-05ba398f25ef\") " pod="openstack/horizon-7bc489768c-rlbc8" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.386399 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/44b58fbc-be53-4531-9dee-05ba398f25ef-horizon-secret-key\") pod \"horizon-7bc489768c-rlbc8\" (UID: \"44b58fbc-be53-4531-9dee-05ba398f25ef\") " pod="openstack/horizon-7bc489768c-rlbc8" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.386434 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/44b58fbc-be53-4531-9dee-05ba398f25ef-scripts\") pod \"horizon-7bc489768c-rlbc8\" (UID: \"44b58fbc-be53-4531-9dee-05ba398f25ef\") " pod="openstack/horizon-7bc489768c-rlbc8" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.386484 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/44b58fbc-be53-4531-9dee-05ba398f25ef-config-data\") pod \"horizon-7bc489768c-rlbc8\" (UID: \"44b58fbc-be53-4531-9dee-05ba398f25ef\") " pod="openstack/horizon-7bc489768c-rlbc8" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.386542 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44b58fbc-be53-4531-9dee-05ba398f25ef-logs\") pod \"horizon-7bc489768c-rlbc8\" (UID: \"44b58fbc-be53-4531-9dee-05ba398f25ef\") " pod="openstack/horizon-7bc489768c-rlbc8" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.386575 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-config-data\") pod \"placement-db-sync-dh4pj\" (UID: \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\") " pod="openstack/placement-db-sync-dh4pj" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.386628 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgm22\" (UniqueName: \"kubernetes.io/projected/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-kube-api-access-sgm22\") pod \"placement-db-sync-dh4pj\" (UID: \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\") " pod="openstack/placement-db-sync-dh4pj" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.386665 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1658991f-7a2b-4ce9-a240-a940385e0b8f-combined-ca-bundle\") pod \"barbican-db-sync-9cmq2\" (UID: \"1658991f-7a2b-4ce9-a240-a940385e0b8f\") " pod="openstack/barbican-db-sync-9cmq2" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.386724 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-scripts\") pod \"placement-db-sync-dh4pj\" (UID: \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\") " pod="openstack/placement-db-sync-dh4pj" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.386748 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-logs\") pod \"placement-db-sync-dh4pj\" (UID: \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\") " pod="openstack/placement-db-sync-dh4pj" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.386805 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-combined-ca-bundle\") pod \"placement-db-sync-dh4pj\" (UID: \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\") " pod="openstack/placement-db-sync-dh4pj" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.388490 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44b58fbc-be53-4531-9dee-05ba398f25ef-logs\") pod \"horizon-7bc489768c-rlbc8\" (UID: \"44b58fbc-be53-4531-9dee-05ba398f25ef\") " pod="openstack/horizon-7bc489768c-rlbc8" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.388678 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/44b58fbc-be53-4531-9dee-05ba398f25ef-scripts\") pod \"horizon-7bc489768c-rlbc8\" (UID: \"44b58fbc-be53-4531-9dee-05ba398f25ef\") " pod="openstack/horizon-7bc489768c-rlbc8" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.390792 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/44b58fbc-be53-4531-9dee-05ba398f25ef-config-data\") pod \"horizon-7bc489768c-rlbc8\" (UID: \"44b58fbc-be53-4531-9dee-05ba398f25ef\") " pod="openstack/horizon-7bc489768c-rlbc8" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.396457 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4nfhc"] Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.399381 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/44b58fbc-be53-4531-9dee-05ba398f25ef-horizon-secret-key\") pod \"horizon-7bc489768c-rlbc8\" (UID: \"44b58fbc-be53-4531-9dee-05ba398f25ef\") " pod="openstack/horizon-7bc489768c-rlbc8" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.403554 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.415866 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" podStartSLOduration=3.415841081 podStartE2EDuration="3.415841081s" podCreationTimestamp="2026-03-21 05:06:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:06:46.251807552 +0000 UTC m=+1159.228271186" watchObservedRunningTime="2026-03-21 05:06:46.415841081 +0000 UTC m=+1159.392304705" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.419839 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tstsp\" (UniqueName: \"kubernetes.io/projected/44b58fbc-be53-4531-9dee-05ba398f25ef-kube-api-access-tstsp\") pod \"horizon-7bc489768c-rlbc8\" (UID: \"44b58fbc-be53-4531-9dee-05ba398f25ef\") " pod="openstack/horizon-7bc489768c-rlbc8" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.450077 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.466503 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bc489768c-rlbc8" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.492763 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-4nfhc\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.492887 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-config-data\") pod \"placement-db-sync-dh4pj\" (UID: \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\") " pod="openstack/placement-db-sync-dh4pj" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.492936 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgm22\" (UniqueName: \"kubernetes.io/projected/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-kube-api-access-sgm22\") pod \"placement-db-sync-dh4pj\" (UID: \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\") " pod="openstack/placement-db-sync-dh4pj" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.492960 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-4nfhc\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.492987 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-4nfhc\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.493025 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1658991f-7a2b-4ce9-a240-a940385e0b8f-combined-ca-bundle\") pod \"barbican-db-sync-9cmq2\" (UID: \"1658991f-7a2b-4ce9-a240-a940385e0b8f\") " pod="openstack/barbican-db-sync-9cmq2" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.493076 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-scripts\") pod \"placement-db-sync-dh4pj\" (UID: \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\") " pod="openstack/placement-db-sync-dh4pj" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.493101 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-logs\") pod \"placement-db-sync-dh4pj\" (UID: \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\") " pod="openstack/placement-db-sync-dh4pj" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.493157 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-combined-ca-bundle\") pod \"placement-db-sync-dh4pj\" (UID: \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\") " pod="openstack/placement-db-sync-dh4pj" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.493266 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-config\") pod \"dnsmasq-dns-cf78879c9-4nfhc\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.493520 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzx7f\" (UniqueName: \"kubernetes.io/projected/aaad9349-8b1e-4f07-b3c9-36bc4781a386-kube-api-access-lzx7f\") pod \"dnsmasq-dns-cf78879c9-4nfhc\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.493590 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1658991f-7a2b-4ce9-a240-a940385e0b8f-db-sync-config-data\") pod \"barbican-db-sync-9cmq2\" (UID: \"1658991f-7a2b-4ce9-a240-a940385e0b8f\") " pod="openstack/barbican-db-sync-9cmq2" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.493639 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-dns-svc\") pod \"dnsmasq-dns-cf78879c9-4nfhc\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.493800 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bczb\" (UniqueName: \"kubernetes.io/projected/1658991f-7a2b-4ce9-a240-a940385e0b8f-kube-api-access-9bczb\") pod \"barbican-db-sync-9cmq2\" (UID: \"1658991f-7a2b-4ce9-a240-a940385e0b8f\") " pod="openstack/barbican-db-sync-9cmq2" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.498201 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-scripts\") pod \"placement-db-sync-dh4pj\" (UID: \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\") " pod="openstack/placement-db-sync-dh4pj" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.500790 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1658991f-7a2b-4ce9-a240-a940385e0b8f-db-sync-config-data\") pod \"barbican-db-sync-9cmq2\" (UID: \"1658991f-7a2b-4ce9-a240-a940385e0b8f\") " pod="openstack/barbican-db-sync-9cmq2" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.502764 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-config-data\") pod \"placement-db-sync-dh4pj\" (UID: \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\") " pod="openstack/placement-db-sync-dh4pj" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.505339 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1658991f-7a2b-4ce9-a240-a940385e0b8f-combined-ca-bundle\") pod \"barbican-db-sync-9cmq2\" (UID: \"1658991f-7a2b-4ce9-a240-a940385e0b8f\") " pod="openstack/barbican-db-sync-9cmq2" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.506083 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-logs\") pod \"placement-db-sync-dh4pj\" (UID: \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\") " pod="openstack/placement-db-sync-dh4pj" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.507017 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-combined-ca-bundle\") pod \"placement-db-sync-dh4pj\" (UID: \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\") " pod="openstack/placement-db-sync-dh4pj" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.521083 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bczb\" (UniqueName: \"kubernetes.io/projected/1658991f-7a2b-4ce9-a240-a940385e0b8f-kube-api-access-9bczb\") pod \"barbican-db-sync-9cmq2\" (UID: \"1658991f-7a2b-4ce9-a240-a940385e0b8f\") " pod="openstack/barbican-db-sync-9cmq2" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.524880 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgm22\" (UniqueName: \"kubernetes.io/projected/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-kube-api-access-sgm22\") pod \"placement-db-sync-dh4pj\" (UID: \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\") " pod="openstack/placement-db-sync-dh4pj" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.580779 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-t7qd7"] Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.595066 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-4nfhc\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.595110 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-4nfhc\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.595207 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-config\") pod \"dnsmasq-dns-cf78879c9-4nfhc\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.595255 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzx7f\" (UniqueName: \"kubernetes.io/projected/aaad9349-8b1e-4f07-b3c9-36bc4781a386-kube-api-access-lzx7f\") pod \"dnsmasq-dns-cf78879c9-4nfhc\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.595294 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-dns-svc\") pod \"dnsmasq-dns-cf78879c9-4nfhc\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.595368 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-4nfhc\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.596445 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-4nfhc\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.597064 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-4nfhc\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.597578 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-4nfhc\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.598149 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-config\") pod \"dnsmasq-dns-cf78879c9-4nfhc\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.598929 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-dns-svc\") pod \"dnsmasq-dns-cf78879c9-4nfhc\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.623021 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzx7f\" (UniqueName: \"kubernetes.io/projected/aaad9349-8b1e-4f07-b3c9-36bc4781a386-kube-api-access-lzx7f\") pod \"dnsmasq-dns-cf78879c9-4nfhc\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.632580 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dh4pj" Mar 21 05:06:46 crc kubenswrapper[4775]: W0321 05:06:46.634495 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded132640_db0b_458b_ba3f_d5aedb284447.slice/crio-80631c4a829e6fbdb4bacbdb17279550e80269d3798a6721d9431adbccaec3b9 WatchSource:0}: Error finding container 80631c4a829e6fbdb4bacbdb17279550e80269d3798a6721d9431adbccaec3b9: Status 404 returned error can't find the container with id 80631c4a829e6fbdb4bacbdb17279550e80269d3798a6721d9431adbccaec3b9 Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.653723 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9cmq2" Mar 21 05:06:46 crc kubenswrapper[4775]: I0321 05:06:46.699195 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.006370 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-9fd475b9c-m48br"] Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.014148 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-wvjlb"] Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.026579 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-kbtmv"] Mar 21 05:06:47 crc kubenswrapper[4775]: W0321 05:06:47.029975 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37bb7e34_ac47_44f6_b18f_ef4ed78eea6a.slice/crio-45cbfcc3ba69b1d165e2a6acc84f20765ed5bc3bad1c098418f5af9089eb0bf1 WatchSource:0}: Error finding container 45cbfcc3ba69b1d165e2a6acc84f20765ed5bc3bad1c098418f5af9089eb0bf1: Status 404 returned error can't find the container with id 45cbfcc3ba69b1d165e2a6acc84f20765ed5bc3bad1c098418f5af9089eb0bf1 Mar 21 05:06:47 crc kubenswrapper[4775]: W0321 05:06:47.031794 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf281a5f0_270c_4d95_93b1_ad265786b886.slice/crio-3ed2947533338d4f3898375bd2e8d20bf0cdf56e90ab25944e2ce0199f079b4a WatchSource:0}: Error finding container 3ed2947533338d4f3898375bd2e8d20bf0cdf56e90ab25944e2ce0199f079b4a: Status 404 returned error can't find the container with id 3ed2947533338d4f3898375bd2e8d20bf0cdf56e90ab25944e2ce0199f079b4a Mar 21 05:06:47 crc kubenswrapper[4775]: W0321 05:06:47.044801 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9531f5f9_8f77_4882_b779_4210b1de81de.slice/crio-a0dba2a88554c5f818ac14151546ad519e642afe620cd4a8580aa68fd7107a3c WatchSource:0}: Error finding container a0dba2a88554c5f818ac14151546ad519e642afe620cd4a8580aa68fd7107a3c: Status 404 returned error can't find the container with id a0dba2a88554c5f818ac14151546ad519e642afe620cd4a8580aa68fd7107a3c Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.050791 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-l8qqc"] Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.187332 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7bc489768c-rlbc8"] Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.244351 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.247728 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9fd475b9c-m48br" event={"ID":"f281a5f0-270c-4d95-93b1-ad265786b886","Type":"ContainerStarted","Data":"3ed2947533338d4f3898375bd2e8d20bf0cdf56e90ab25944e2ce0199f079b4a"} Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.249592 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-l8qqc" event={"ID":"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a","Type":"ContainerStarted","Data":"45cbfcc3ba69b1d165e2a6acc84f20765ed5bc3bad1c098418f5af9089eb0bf1"} Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.251178 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wvjlb" event={"ID":"9531f5f9-8f77-4882-b779-4210b1de81de","Type":"ContainerStarted","Data":"a0dba2a88554c5f818ac14151546ad519e642afe620cd4a8580aa68fd7107a3c"} Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.252587 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-t7qd7" event={"ID":"ed132640-db0b-458b-ba3f-d5aedb284447","Type":"ContainerStarted","Data":"80631c4a829e6fbdb4bacbdb17279550e80269d3798a6721d9431adbccaec3b9"} Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.259686 4775 generic.go:334] "Generic (PLEG): container finished" podID="e2c4bdcc-f0d3-452d-87ac-52ce486f1c52" containerID="44277cb33149a33c5c08b0bf964d1c1e6f322cfb36b55ecaba7cde47956ea895" exitCode=0 Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.259825 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" event={"ID":"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52","Type":"ContainerDied","Data":"44277cb33149a33c5c08b0bf964d1c1e6f322cfb36b55ecaba7cde47956ea895"} Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.279857 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kbtmv" event={"ID":"50f0ec15-decb-4879-9410-0962a21f83ae","Type":"ContainerStarted","Data":"241678b559ba9174070a5793922c2b069fb839b133fbc00760a37f9bf9777ed9"} Mar 21 05:06:47 crc kubenswrapper[4775]: W0321 05:06:47.435706 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cd9d81f_8c8a_46f9_9943_42dc0b638bef.slice/crio-f10dda8724855df58c36f76f7e138a9aa8f4ab4f82a8c7d9fc8bb963f6bad44c WatchSource:0}: Error finding container f10dda8724855df58c36f76f7e138a9aa8f4ab4f82a8c7d9fc8bb963f6bad44c: Status 404 returned error can't find the container with id f10dda8724855df58c36f76f7e138a9aa8f4ab4f82a8c7d9fc8bb963f6bad44c Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.448893 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-dh4pj"] Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.486243 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9cmq2"] Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.514962 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7bc489768c-rlbc8"] Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.527585 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6c4c5b5499-jnmhl"] Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.529376 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c4c5b5499-jnmhl" Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.534088 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6c4c5b5499-jnmhl"] Mar 21 05:06:47 crc kubenswrapper[4775]: W0321 05:06:47.547084 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaaad9349_8b1e_4f07_b3c9_36bc4781a386.slice/crio-545d0a4dfbecffb0f9c29e4cf099e382302c9c5682c08fe8a7fbe8e6c0247a35 WatchSource:0}: Error finding container 545d0a4dfbecffb0f9c29e4cf099e382302c9c5682c08fe8a7fbe8e6c0247a35: Status 404 returned error can't find the container with id 545d0a4dfbecffb0f9c29e4cf099e382302c9c5682c08fe8a7fbe8e6c0247a35 Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.567713 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4nfhc"] Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.572975 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.588891 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.732483 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-dns-swift-storage-0\") pod \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.732669 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-ovsdbserver-nb\") pod \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.732707 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-config\") pod \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.732782 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-dns-svc\") pod \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.732861 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s69ql\" (UniqueName: \"kubernetes.io/projected/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-kube-api-access-s69ql\") pod \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.733030 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-ovsdbserver-sb\") pod \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\" (UID: \"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52\") " Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.733460 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbdnp\" (UniqueName: \"kubernetes.io/projected/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-kube-api-access-rbdnp\") pod \"horizon-6c4c5b5499-jnmhl\" (UID: \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\") " pod="openstack/horizon-6c4c5b5499-jnmhl" Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.733568 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-scripts\") pod \"horizon-6c4c5b5499-jnmhl\" (UID: \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\") " pod="openstack/horizon-6c4c5b5499-jnmhl" Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.733632 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-logs\") pod \"horizon-6c4c5b5499-jnmhl\" (UID: \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\") " pod="openstack/horizon-6c4c5b5499-jnmhl" Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.733981 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-config-data\") pod \"horizon-6c4c5b5499-jnmhl\" (UID: \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\") " pod="openstack/horizon-6c4c5b5499-jnmhl" Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.734026 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-horizon-secret-key\") pod \"horizon-6c4c5b5499-jnmhl\" (UID: \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\") " pod="openstack/horizon-6c4c5b5499-jnmhl" Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.758163 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-kube-api-access-s69ql" (OuterVolumeSpecName: "kube-api-access-s69ql") pod "e2c4bdcc-f0d3-452d-87ac-52ce486f1c52" (UID: "e2c4bdcc-f0d3-452d-87ac-52ce486f1c52"). InnerVolumeSpecName "kube-api-access-s69ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.835416 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-config-data\") pod \"horizon-6c4c5b5499-jnmhl\" (UID: \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\") " pod="openstack/horizon-6c4c5b5499-jnmhl" Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.841548 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-config-data\") pod \"horizon-6c4c5b5499-jnmhl\" (UID: \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\") " pod="openstack/horizon-6c4c5b5499-jnmhl" Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.841696 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-horizon-secret-key\") pod \"horizon-6c4c5b5499-jnmhl\" (UID: \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\") " pod="openstack/horizon-6c4c5b5499-jnmhl" Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.842468 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbdnp\" (UniqueName: \"kubernetes.io/projected/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-kube-api-access-rbdnp\") pod \"horizon-6c4c5b5499-jnmhl\" (UID: \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\") " pod="openstack/horizon-6c4c5b5499-jnmhl" Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.842642 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-scripts\") pod \"horizon-6c4c5b5499-jnmhl\" (UID: \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\") " pod="openstack/horizon-6c4c5b5499-jnmhl" Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.842700 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-logs\") pod \"horizon-6c4c5b5499-jnmhl\" (UID: \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\") " pod="openstack/horizon-6c4c5b5499-jnmhl" Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.842975 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s69ql\" (UniqueName: \"kubernetes.io/projected/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-kube-api-access-s69ql\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.843798 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-scripts\") pod \"horizon-6c4c5b5499-jnmhl\" (UID: \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\") " pod="openstack/horizon-6c4c5b5499-jnmhl" Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.843801 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-logs\") pod \"horizon-6c4c5b5499-jnmhl\" (UID: \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\") " pod="openstack/horizon-6c4c5b5499-jnmhl" Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.849345 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-horizon-secret-key\") pod \"horizon-6c4c5b5499-jnmhl\" (UID: \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\") " pod="openstack/horizon-6c4c5b5499-jnmhl" Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.880754 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbdnp\" (UniqueName: \"kubernetes.io/projected/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-kube-api-access-rbdnp\") pod \"horizon-6c4c5b5499-jnmhl\" (UID: \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\") " pod="openstack/horizon-6c4c5b5499-jnmhl" Mar 21 05:06:47 crc kubenswrapper[4775]: I0321 05:06:47.898025 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c4c5b5499-jnmhl" Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.042515 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e2c4bdcc-f0d3-452d-87ac-52ce486f1c52" (UID: "e2c4bdcc-f0d3-452d-87ac-52ce486f1c52"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.048790 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-config" (OuterVolumeSpecName: "config") pod "e2c4bdcc-f0d3-452d-87ac-52ce486f1c52" (UID: "e2c4bdcc-f0d3-452d-87ac-52ce486f1c52"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.049916 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.049931 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.052670 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e2c4bdcc-f0d3-452d-87ac-52ce486f1c52" (UID: "e2c4bdcc-f0d3-452d-87ac-52ce486f1c52"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.052837 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e2c4bdcc-f0d3-452d-87ac-52ce486f1c52" (UID: "e2c4bdcc-f0d3-452d-87ac-52ce486f1c52"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.062605 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e2c4bdcc-f0d3-452d-87ac-52ce486f1c52" (UID: "e2c4bdcc-f0d3-452d-87ac-52ce486f1c52"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.151178 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.151219 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.151230 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.308628 4775 generic.go:334] "Generic (PLEG): container finished" podID="ed132640-db0b-458b-ba3f-d5aedb284447" containerID="0b3ac32d8cf0aa92220dc7dc36f6c6433239ed9b4e495f670922d65ddd3a5492" exitCode=0 Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.308917 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-t7qd7" event={"ID":"ed132640-db0b-458b-ba3f-d5aedb284447","Type":"ContainerDied","Data":"0b3ac32d8cf0aa92220dc7dc36f6c6433239ed9b4e495f670922d65ddd3a5492"} Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.314304 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dh4pj" event={"ID":"5cd9d81f-8c8a-46f9-9943-42dc0b638bef","Type":"ContainerStarted","Data":"f10dda8724855df58c36f76f7e138a9aa8f4ab4f82a8c7d9fc8bb963f6bad44c"} Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.317736 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" event={"ID":"e2c4bdcc-f0d3-452d-87ac-52ce486f1c52","Type":"ContainerDied","Data":"ddb13a088c51c9a2b5faad8f81230304cbd7e86c02e650cace18f83a00dc10db"} Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.317794 4775 scope.go:117] "RemoveContainer" containerID="44277cb33149a33c5c08b0bf964d1c1e6f322cfb36b55ecaba7cde47956ea895" Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.317914 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-jhdxv" Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.333630 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wvjlb" event={"ID":"9531f5f9-8f77-4882-b779-4210b1de81de","Type":"ContainerStarted","Data":"f19a3eb592d811463bd26fefd306c16b1ec6187913716e5585ce5b02489e2c0c"} Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.341013 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9cmq2" event={"ID":"1658991f-7a2b-4ce9-a240-a940385e0b8f","Type":"ContainerStarted","Data":"cb42616105e59d6830a168b6da9f0c076bcf19cd878442d47ceb6cd33726aca0"} Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.342754 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kbtmv" event={"ID":"50f0ec15-decb-4879-9410-0962a21f83ae","Type":"ContainerStarted","Data":"fdcd202238496f57430b4ad2beaac05c4b222263df155d2d3bab8e200bbd88c5"} Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.345339 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bc489768c-rlbc8" event={"ID":"44b58fbc-be53-4531-9dee-05ba398f25ef","Type":"ContainerStarted","Data":"e60f7afbe17b7aa1bbfaf8e491a43aeb53519de12ebcef7dcaf29708ec522108"} Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.354021 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ac5c16c-56ac-4299-ae61-a8200986ce10","Type":"ContainerStarted","Data":"25ef4b17ffe6969bdd8510a4d651e67244e698c334397d84b7e113488d96c6fb"} Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.365694 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-wvjlb" podStartSLOduration=3.365671423 podStartE2EDuration="3.365671423s" podCreationTimestamp="2026-03-21 05:06:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:06:48.349219568 +0000 UTC m=+1161.325683202" watchObservedRunningTime="2026-03-21 05:06:48.365671423 +0000 UTC m=+1161.342135047" Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.374611 4775 scope.go:117] "RemoveContainer" containerID="11cb266c6b3aa03b4030e08a2478a4c8badfb6b932b90e369c8e339bd2194478" Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.374673 4775 generic.go:334] "Generic (PLEG): container finished" podID="aaad9349-8b1e-4f07-b3c9-36bc4781a386" containerID="4bd42a0432eaefd26033f0ffdb594981243afbef4a05205089435812a0be336b" exitCode=0 Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.374722 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" event={"ID":"aaad9349-8b1e-4f07-b3c9-36bc4781a386","Type":"ContainerDied","Data":"4bd42a0432eaefd26033f0ffdb594981243afbef4a05205089435812a0be336b"} Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.374754 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" event={"ID":"aaad9349-8b1e-4f07-b3c9-36bc4781a386","Type":"ContainerStarted","Data":"545d0a4dfbecffb0f9c29e4cf099e382302c9c5682c08fe8a7fbe8e6c0247a35"} Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.384160 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-jhdxv"] Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.391373 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-jhdxv"] Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.399447 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-kbtmv" podStartSLOduration=3.399434588 podStartE2EDuration="3.399434588s" podCreationTimestamp="2026-03-21 05:06:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:06:48.396351971 +0000 UTC m=+1161.372815605" watchObservedRunningTime="2026-03-21 05:06:48.399434588 +0000 UTC m=+1161.375898212" Mar 21 05:06:48 crc kubenswrapper[4775]: I0321 05:06:48.480361 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6c4c5b5499-jnmhl"] Mar 21 05:06:49 crc kubenswrapper[4775]: I0321 05:06:49.674661 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2c4bdcc-f0d3-452d-87ac-52ce486f1c52" path="/var/lib/kubelet/pods/e2c4bdcc-f0d3-452d-87ac-52ce486f1c52/volumes" Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.404268 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c4c5b5499-jnmhl" event={"ID":"e7fbefc1-c1f6-472d-82b4-77c0a44b514f","Type":"ContainerStarted","Data":"4d928750acf27a12d43a71655b2e716129e9bfb358dbdddb74f0ae57f062cbba"} Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.405296 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.406080 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-t7qd7" event={"ID":"ed132640-db0b-458b-ba3f-d5aedb284447","Type":"ContainerDied","Data":"80631c4a829e6fbdb4bacbdb17279550e80269d3798a6721d9431adbccaec3b9"} Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.406147 4775 scope.go:117] "RemoveContainer" containerID="0b3ac32d8cf0aa92220dc7dc36f6c6433239ed9b4e495f670922d65ddd3a5492" Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.519501 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-ovsdbserver-nb\") pod \"ed132640-db0b-458b-ba3f-d5aedb284447\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.519621 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-config\") pod \"ed132640-db0b-458b-ba3f-d5aedb284447\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.519679 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7m66r\" (UniqueName: \"kubernetes.io/projected/ed132640-db0b-458b-ba3f-d5aedb284447-kube-api-access-7m66r\") pod \"ed132640-db0b-458b-ba3f-d5aedb284447\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.519731 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-dns-svc\") pod \"ed132640-db0b-458b-ba3f-d5aedb284447\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.519790 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-dns-swift-storage-0\") pod \"ed132640-db0b-458b-ba3f-d5aedb284447\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.519832 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-ovsdbserver-sb\") pod \"ed132640-db0b-458b-ba3f-d5aedb284447\" (UID: \"ed132640-db0b-458b-ba3f-d5aedb284447\") " Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.526975 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed132640-db0b-458b-ba3f-d5aedb284447-kube-api-access-7m66r" (OuterVolumeSpecName: "kube-api-access-7m66r") pod "ed132640-db0b-458b-ba3f-d5aedb284447" (UID: "ed132640-db0b-458b-ba3f-d5aedb284447"). InnerVolumeSpecName "kube-api-access-7m66r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.546849 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ed132640-db0b-458b-ba3f-d5aedb284447" (UID: "ed132640-db0b-458b-ba3f-d5aedb284447"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.550562 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ed132640-db0b-458b-ba3f-d5aedb284447" (UID: "ed132640-db0b-458b-ba3f-d5aedb284447"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.553637 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ed132640-db0b-458b-ba3f-d5aedb284447" (UID: "ed132640-db0b-458b-ba3f-d5aedb284447"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.554576 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ed132640-db0b-458b-ba3f-d5aedb284447" (UID: "ed132640-db0b-458b-ba3f-d5aedb284447"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.561306 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-config" (OuterVolumeSpecName: "config") pod "ed132640-db0b-458b-ba3f-d5aedb284447" (UID: "ed132640-db0b-458b-ba3f-d5aedb284447"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.621655 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.621699 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7m66r\" (UniqueName: \"kubernetes.io/projected/ed132640-db0b-458b-ba3f-d5aedb284447-kube-api-access-7m66r\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.621714 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.621729 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.621744 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:50 crc kubenswrapper[4775]: I0321 05:06:50.621756 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed132640-db0b-458b-ba3f-d5aedb284447-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 05:06:51 crc kubenswrapper[4775]: I0321 05:06:51.456231 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-td76j" event={"ID":"716605f1-5111-4e7a-9591-18dfb5da1984","Type":"ContainerStarted","Data":"43def41a0cf1f318d43db6319e03e84d1a862d118a43d560e7c5e1d51f26dead"} Mar 21 05:06:51 crc kubenswrapper[4775]: I0321 05:06:51.462337 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-t7qd7" Mar 21 05:06:51 crc kubenswrapper[4775]: I0321 05:06:51.465171 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" event={"ID":"aaad9349-8b1e-4f07-b3c9-36bc4781a386","Type":"ContainerStarted","Data":"f2fafd7848327783c5d640051878058b230f83637eec544e85f17df14ee71f99"} Mar 21 05:06:51 crc kubenswrapper[4775]: I0321 05:06:51.465757 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:51 crc kubenswrapper[4775]: I0321 05:06:51.476754 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-td76j" podStartSLOduration=2.708029554 podStartE2EDuration="39.476733655s" podCreationTimestamp="2026-03-21 05:06:12 +0000 UTC" firstStartedPulling="2026-03-21 05:06:13.503500319 +0000 UTC m=+1126.479963943" lastFinishedPulling="2026-03-21 05:06:50.27220442 +0000 UTC m=+1163.248668044" observedRunningTime="2026-03-21 05:06:51.472502625 +0000 UTC m=+1164.448966249" watchObservedRunningTime="2026-03-21 05:06:51.476733655 +0000 UTC m=+1164.453197279" Mar 21 05:06:51 crc kubenswrapper[4775]: I0321 05:06:51.504480 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" podStartSLOduration=5.504459889 podStartE2EDuration="5.504459889s" podCreationTimestamp="2026-03-21 05:06:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:06:51.491458161 +0000 UTC m=+1164.467921795" watchObservedRunningTime="2026-03-21 05:06:51.504459889 +0000 UTC m=+1164.480923513" Mar 21 05:06:51 crc kubenswrapper[4775]: I0321 05:06:51.561587 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-t7qd7"] Mar 21 05:06:51 crc kubenswrapper[4775]: I0321 05:06:51.575830 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-t7qd7"] Mar 21 05:06:51 crc kubenswrapper[4775]: I0321 05:06:51.693267 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed132640-db0b-458b-ba3f-d5aedb284447" path="/var/lib/kubelet/pods/ed132640-db0b-458b-ba3f-d5aedb284447/volumes" Mar 21 05:06:53 crc kubenswrapper[4775]: I0321 05:06:53.483921 4775 generic.go:334] "Generic (PLEG): container finished" podID="50f0ec15-decb-4879-9410-0962a21f83ae" containerID="fdcd202238496f57430b4ad2beaac05c4b222263df155d2d3bab8e200bbd88c5" exitCode=0 Mar 21 05:06:53 crc kubenswrapper[4775]: I0321 05:06:53.483971 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kbtmv" event={"ID":"50f0ec15-decb-4879-9410-0962a21f83ae","Type":"ContainerDied","Data":"fdcd202238496f57430b4ad2beaac05c4b222263df155d2d3bab8e200bbd88c5"} Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.572960 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-9fd475b9c-m48br"] Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.634085 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-64fb567758-hd2ld"] Mar 21 05:06:54 crc kubenswrapper[4775]: E0321 05:06:54.634880 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed132640-db0b-458b-ba3f-d5aedb284447" containerName="init" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.634958 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed132640-db0b-458b-ba3f-d5aedb284447" containerName="init" Mar 21 05:06:54 crc kubenswrapper[4775]: E0321 05:06:54.635027 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2c4bdcc-f0d3-452d-87ac-52ce486f1c52" containerName="init" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.637474 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2c4bdcc-f0d3-452d-87ac-52ce486f1c52" containerName="init" Mar 21 05:06:54 crc kubenswrapper[4775]: E0321 05:06:54.637645 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2c4bdcc-f0d3-452d-87ac-52ce486f1c52" containerName="dnsmasq-dns" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.637722 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2c4bdcc-f0d3-452d-87ac-52ce486f1c52" containerName="dnsmasq-dns" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.638044 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2c4bdcc-f0d3-452d-87ac-52ce486f1c52" containerName="dnsmasq-dns" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.638109 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed132640-db0b-458b-ba3f-d5aedb284447" containerName="init" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.639419 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.643437 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.664313 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64fb567758-hd2ld"] Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.679832 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6c4c5b5499-jnmhl"] Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.700344 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6496ddbdd4-v5mc5"] Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.701680 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.721914 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6496ddbdd4-v5mc5"] Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.731795 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdmrs\" (UniqueName: \"kubernetes.io/projected/fc6e433f-9e70-4b09-9780-403634bbe0dc-kube-api-access-mdmrs\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.732091 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09cf3763-0b41-4452-a247-d9a56f58b05d-config-data\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.732218 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fc6e433f-9e70-4b09-9780-403634bbe0dc-horizon-secret-key\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.732320 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc6e433f-9e70-4b09-9780-403634bbe0dc-logs\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.732439 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09cf3763-0b41-4452-a247-d9a56f58b05d-logs\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.732532 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fc6e433f-9e70-4b09-9780-403634bbe0dc-config-data\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.732616 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/09cf3763-0b41-4452-a247-d9a56f58b05d-horizon-tls-certs\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.732699 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc6e433f-9e70-4b09-9780-403634bbe0dc-combined-ca-bundle\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.732823 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc6e433f-9e70-4b09-9780-403634bbe0dc-scripts\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.732918 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09cf3763-0b41-4452-a247-d9a56f58b05d-combined-ca-bundle\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.733004 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc6e433f-9e70-4b09-9780-403634bbe0dc-horizon-tls-certs\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.733162 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09cf3763-0b41-4452-a247-d9a56f58b05d-scripts\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.733272 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg4gg\" (UniqueName: \"kubernetes.io/projected/09cf3763-0b41-4452-a247-d9a56f58b05d-kube-api-access-rg4gg\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.733371 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/09cf3763-0b41-4452-a247-d9a56f58b05d-horizon-secret-key\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.833967 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09cf3763-0b41-4452-a247-d9a56f58b05d-logs\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.834041 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/09cf3763-0b41-4452-a247-d9a56f58b05d-horizon-tls-certs\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.834074 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fc6e433f-9e70-4b09-9780-403634bbe0dc-config-data\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.834096 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc6e433f-9e70-4b09-9780-403634bbe0dc-combined-ca-bundle\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.834166 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc6e433f-9e70-4b09-9780-403634bbe0dc-scripts\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.834189 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09cf3763-0b41-4452-a247-d9a56f58b05d-combined-ca-bundle\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.834244 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc6e433f-9e70-4b09-9780-403634bbe0dc-horizon-tls-certs\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.834329 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09cf3763-0b41-4452-a247-d9a56f58b05d-scripts\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.834365 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg4gg\" (UniqueName: \"kubernetes.io/projected/09cf3763-0b41-4452-a247-d9a56f58b05d-kube-api-access-rg4gg\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.834416 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/09cf3763-0b41-4452-a247-d9a56f58b05d-horizon-secret-key\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.834445 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdmrs\" (UniqueName: \"kubernetes.io/projected/fc6e433f-9e70-4b09-9780-403634bbe0dc-kube-api-access-mdmrs\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.834498 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09cf3763-0b41-4452-a247-d9a56f58b05d-config-data\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.834522 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fc6e433f-9e70-4b09-9780-403634bbe0dc-horizon-secret-key\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.834560 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc6e433f-9e70-4b09-9780-403634bbe0dc-logs\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.834708 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09cf3763-0b41-4452-a247-d9a56f58b05d-logs\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.835198 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc6e433f-9e70-4b09-9780-403634bbe0dc-logs\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.836255 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fc6e433f-9e70-4b09-9780-403634bbe0dc-config-data\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.836947 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09cf3763-0b41-4452-a247-d9a56f58b05d-scripts\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.837066 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc6e433f-9e70-4b09-9780-403634bbe0dc-scripts\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.837297 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09cf3763-0b41-4452-a247-d9a56f58b05d-config-data\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.843072 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fc6e433f-9e70-4b09-9780-403634bbe0dc-horizon-secret-key\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.844193 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/09cf3763-0b41-4452-a247-d9a56f58b05d-horizon-secret-key\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.849212 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09cf3763-0b41-4452-a247-d9a56f58b05d-combined-ca-bundle\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.853336 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc6e433f-9e70-4b09-9780-403634bbe0dc-combined-ca-bundle\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.853751 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc6e433f-9e70-4b09-9780-403634bbe0dc-horizon-tls-certs\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.857423 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/09cf3763-0b41-4452-a247-d9a56f58b05d-horizon-tls-certs\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.857874 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg4gg\" (UniqueName: \"kubernetes.io/projected/09cf3763-0b41-4452-a247-d9a56f58b05d-kube-api-access-rg4gg\") pod \"horizon-64fb567758-hd2ld\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.861972 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdmrs\" (UniqueName: \"kubernetes.io/projected/fc6e433f-9e70-4b09-9780-403634bbe0dc-kube-api-access-mdmrs\") pod \"horizon-6496ddbdd4-v5mc5\" (UID: \"fc6e433f-9e70-4b09-9780-403634bbe0dc\") " pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:54 crc kubenswrapper[4775]: I0321 05:06:54.978646 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:06:55 crc kubenswrapper[4775]: I0321 05:06:55.031531 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:06:56 crc kubenswrapper[4775]: I0321 05:06:56.708202 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:06:56 crc kubenswrapper[4775]: I0321 05:06:56.767552 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-b4m98"] Mar 21 05:06:56 crc kubenswrapper[4775]: I0321 05:06:56.768193 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" podUID="6b50afbd-31b8-40ff-bd7b-1ce5021e2837" containerName="dnsmasq-dns" containerID="cri-o://6c721bd1cf19d6ad1fc2eb055e74ccf9afea13852ad17eeb02e87ea598c42465" gracePeriod=10 Mar 21 05:06:58 crc kubenswrapper[4775]: I0321 05:06:58.544644 4775 generic.go:334] "Generic (PLEG): container finished" podID="6b50afbd-31b8-40ff-bd7b-1ce5021e2837" containerID="6c721bd1cf19d6ad1fc2eb055e74ccf9afea13852ad17eeb02e87ea598c42465" exitCode=0 Mar 21 05:06:58 crc kubenswrapper[4775]: I0321 05:06:58.544759 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" event={"ID":"6b50afbd-31b8-40ff-bd7b-1ce5021e2837","Type":"ContainerDied","Data":"6c721bd1cf19d6ad1fc2eb055e74ccf9afea13852ad17eeb02e87ea598c42465"} Mar 21 05:07:01 crc kubenswrapper[4775]: I0321 05:07:01.095021 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" podUID="6b50afbd-31b8-40ff-bd7b-1ce5021e2837" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: connect: connection refused" Mar 21 05:07:02 crc kubenswrapper[4775]: I0321 05:07:02.482870 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:07:02 crc kubenswrapper[4775]: I0321 05:07:02.482946 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:07:06 crc kubenswrapper[4775]: I0321 05:07:06.095513 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" podUID="6b50afbd-31b8-40ff-bd7b-1ce5021e2837" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: connect: connection refused" Mar 21 05:07:06 crc kubenswrapper[4775]: E0321 05:07:06.719655 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Mar 21 05:07:06 crc kubenswrapper[4775]: E0321 05:07:06.720313 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5h5bbh9bh666h594h66fhcch68bh5dch9bh649hcch5fh9dh676h79h698h79h687h8hf7h74h5dh57h6ch655h5b4h86h554h664hdch647q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jjzgj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-9fd475b9c-m48br_openstack(f281a5f0-270c-4d95-93b1-ad265786b886): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 05:07:06 crc kubenswrapper[4775]: E0321 05:07:06.722983 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-9fd475b9c-m48br" podUID="f281a5f0-270c-4d95-93b1-ad265786b886" Mar 21 05:07:07 crc kubenswrapper[4775]: E0321 05:07:07.024156 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Mar 21 05:07:07 crc kubenswrapper[4775]: E0321 05:07:07.024353 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h568h666hcfh5ch64bh67ch75h96h645h697h66fh58fh56fh548hf9h554h5c8h669h588h75hbbh64dh596h689h585hb5h68dh87h579h69h5q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w8h8k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(8ac5c16c-56ac-4299-ae61-a8200986ce10): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 05:07:16 crc kubenswrapper[4775]: I0321 05:07:16.094846 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" podUID="6b50afbd-31b8-40ff-bd7b-1ce5021e2837" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: i/o timeout" Mar 21 05:07:16 crc kubenswrapper[4775]: I0321 05:07:16.095561 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.095821 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" podUID="6b50afbd-31b8-40ff-bd7b-1ce5021e2837" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: i/o timeout" Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.791663 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kbtmv" event={"ID":"50f0ec15-decb-4879-9410-0962a21f83ae","Type":"ContainerDied","Data":"241678b559ba9174070a5793922c2b069fb839b133fbc00760a37f9bf9777ed9"} Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.791927 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="241678b559ba9174070a5793922c2b069fb839b133fbc00760a37f9bf9777ed9" Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.803427 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.887855 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-combined-ca-bundle\") pod \"50f0ec15-decb-4879-9410-0962a21f83ae\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.887929 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-fernet-keys\") pod \"50f0ec15-decb-4879-9410-0962a21f83ae\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.888012 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-credential-keys\") pod \"50f0ec15-decb-4879-9410-0962a21f83ae\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.888047 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-scripts\") pod \"50f0ec15-decb-4879-9410-0962a21f83ae\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.888108 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-config-data\") pod \"50f0ec15-decb-4879-9410-0962a21f83ae\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.888157 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znghz\" (UniqueName: \"kubernetes.io/projected/50f0ec15-decb-4879-9410-0962a21f83ae-kube-api-access-znghz\") pod \"50f0ec15-decb-4879-9410-0962a21f83ae\" (UID: \"50f0ec15-decb-4879-9410-0962a21f83ae\") " Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.894539 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "50f0ec15-decb-4879-9410-0962a21f83ae" (UID: "50f0ec15-decb-4879-9410-0962a21f83ae"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.894788 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-scripts" (OuterVolumeSpecName: "scripts") pod "50f0ec15-decb-4879-9410-0962a21f83ae" (UID: "50f0ec15-decb-4879-9410-0962a21f83ae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.895450 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50f0ec15-decb-4879-9410-0962a21f83ae-kube-api-access-znghz" (OuterVolumeSpecName: "kube-api-access-znghz") pod "50f0ec15-decb-4879-9410-0962a21f83ae" (UID: "50f0ec15-decb-4879-9410-0962a21f83ae"). InnerVolumeSpecName "kube-api-access-znghz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.898302 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "50f0ec15-decb-4879-9410-0962a21f83ae" (UID: "50f0ec15-decb-4879-9410-0962a21f83ae"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.918789 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50f0ec15-decb-4879-9410-0962a21f83ae" (UID: "50f0ec15-decb-4879-9410-0962a21f83ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.936850 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-config-data" (OuterVolumeSpecName: "config-data") pod "50f0ec15-decb-4879-9410-0962a21f83ae" (UID: "50f0ec15-decb-4879-9410-0962a21f83ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.989834 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znghz\" (UniqueName: \"kubernetes.io/projected/50f0ec15-decb-4879-9410-0962a21f83ae-kube-api-access-znghz\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.989865 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.989875 4775 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.989883 4775 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.989892 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:21 crc kubenswrapper[4775]: I0321 05:07:21.989900 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50f0ec15-decb-4879-9410-0962a21f83ae-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:22 crc kubenswrapper[4775]: I0321 05:07:22.800194 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kbtmv" Mar 21 05:07:22 crc kubenswrapper[4775]: I0321 05:07:22.898130 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-kbtmv"] Mar 21 05:07:22 crc kubenswrapper[4775]: I0321 05:07:22.906312 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-kbtmv"] Mar 21 05:07:22 crc kubenswrapper[4775]: I0321 05:07:22.994614 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-jxmtf"] Mar 21 05:07:22 crc kubenswrapper[4775]: E0321 05:07:22.994974 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50f0ec15-decb-4879-9410-0962a21f83ae" containerName="keystone-bootstrap" Mar 21 05:07:22 crc kubenswrapper[4775]: I0321 05:07:22.994986 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="50f0ec15-decb-4879-9410-0962a21f83ae" containerName="keystone-bootstrap" Mar 21 05:07:22 crc kubenswrapper[4775]: I0321 05:07:22.995952 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="50f0ec15-decb-4879-9410-0962a21f83ae" containerName="keystone-bootstrap" Mar 21 05:07:22 crc kubenswrapper[4775]: I0321 05:07:22.996563 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.001650 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.001862 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.002002 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-nhwlw" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.003443 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.007507 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.013569 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jxmtf"] Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.108372 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx7pf\" (UniqueName: \"kubernetes.io/projected/cdf8b79b-6156-4aa5-a769-3a96408745c1-kube-api-access-dx7pf\") pod \"keystone-bootstrap-jxmtf\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.108471 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-config-data\") pod \"keystone-bootstrap-jxmtf\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.108503 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-combined-ca-bundle\") pod \"keystone-bootstrap-jxmtf\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.108550 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-scripts\") pod \"keystone-bootstrap-jxmtf\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.108567 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-fernet-keys\") pod \"keystone-bootstrap-jxmtf\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.108801 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-credential-keys\") pod \"keystone-bootstrap-jxmtf\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.210984 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-scripts\") pod \"keystone-bootstrap-jxmtf\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.211059 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-fernet-keys\") pod \"keystone-bootstrap-jxmtf\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.211664 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-credential-keys\") pod \"keystone-bootstrap-jxmtf\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.211735 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx7pf\" (UniqueName: \"kubernetes.io/projected/cdf8b79b-6156-4aa5-a769-3a96408745c1-kube-api-access-dx7pf\") pod \"keystone-bootstrap-jxmtf\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.211808 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-config-data\") pod \"keystone-bootstrap-jxmtf\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.211925 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-combined-ca-bundle\") pod \"keystone-bootstrap-jxmtf\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.216415 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-credential-keys\") pod \"keystone-bootstrap-jxmtf\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.216916 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-scripts\") pod \"keystone-bootstrap-jxmtf\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.217046 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-fernet-keys\") pod \"keystone-bootstrap-jxmtf\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.219794 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-config-data\") pod \"keystone-bootstrap-jxmtf\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.221003 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-combined-ca-bundle\") pod \"keystone-bootstrap-jxmtf\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.228255 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx7pf\" (UniqueName: \"kubernetes.io/projected/cdf8b79b-6156-4aa5-a769-3a96408745c1-kube-api-access-dx7pf\") pod \"keystone-bootstrap-jxmtf\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.318167 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:23 crc kubenswrapper[4775]: I0321 05:07:23.673606 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50f0ec15-decb-4879-9410-0962a21f83ae" path="/var/lib/kubelet/pods/50f0ec15-decb-4879-9410-0962a21f83ae/volumes" Mar 21 05:07:23 crc kubenswrapper[4775]: E0321 05:07:23.771483 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Mar 21 05:07:23 crc kubenswrapper[4775]: E0321 05:07:23.771641 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sgm22,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-dh4pj_openstack(5cd9d81f-8c8a-46f9-9943-42dc0b638bef): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 05:07:23 crc kubenswrapper[4775]: E0321 05:07:23.773035 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-dh4pj" podUID="5cd9d81f-8c8a-46f9-9943-42dc0b638bef" Mar 21 05:07:23 crc kubenswrapper[4775]: E0321 05:07:23.809313 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-dh4pj" podUID="5cd9d81f-8c8a-46f9-9943-42dc0b638bef" Mar 21 05:07:24 crc kubenswrapper[4775]: E0321 05:07:24.865680 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Mar 21 05:07:24 crc kubenswrapper[4775]: E0321 05:07:24.866146 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nf9h5c5h5d9h99h559h585h84h5fch7fh584h589h64h5fch695h5f5h65fh5h88h564h645h547hcfhb5hfdh55fh596h8ch649hd4h577h65fh599q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rbdnp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6c4c5b5499-jnmhl_openstack(e7fbefc1-c1f6-472d-82b4-77c0a44b514f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 05:07:24 crc kubenswrapper[4775]: E0321 05:07:24.870168 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-6c4c5b5499-jnmhl" podUID="e7fbefc1-c1f6-472d-82b4-77c0a44b514f" Mar 21 05:07:24 crc kubenswrapper[4775]: E0321 05:07:24.870815 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = reading blob sha256:5ee16af83557284a12c5915ee8d5d1cc2bc45e245e379412416cadf7824031aa: Get \"https://quay.io/v2/podified-antelope-centos9/openstack-barbican-api/blobs/sha256:5ee16af83557284a12c5915ee8d5d1cc2bc45e245e379412416cadf7824031aa\": context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Mar 21 05:07:24 crc kubenswrapper[4775]: E0321 05:07:24.871029 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9bczb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-9cmq2_openstack(1658991f-7a2b-4ce9-a240-a940385e0b8f): ErrImagePull: rpc error: code = Canceled desc = reading blob sha256:5ee16af83557284a12c5915ee8d5d1cc2bc45e245e379412416cadf7824031aa: Get \"https://quay.io/v2/podified-antelope-centos9/openstack-barbican-api/blobs/sha256:5ee16af83557284a12c5915ee8d5d1cc2bc45e245e379412416cadf7824031aa\": context canceled" logger="UnhandledError" Mar 21 05:07:24 crc kubenswrapper[4775]: E0321 05:07:24.873022 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = reading blob sha256:5ee16af83557284a12c5915ee8d5d1cc2bc45e245e379412416cadf7824031aa: Get \\\"https://quay.io/v2/podified-antelope-centos9/openstack-barbican-api/blobs/sha256:5ee16af83557284a12c5915ee8d5d1cc2bc45e245e379412416cadf7824031aa\\\": context canceled\"" pod="openstack/barbican-db-sync-9cmq2" podUID="1658991f-7a2b-4ce9-a240-a940385e0b8f" Mar 21 05:07:24 crc kubenswrapper[4775]: E0321 05:07:24.884574 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Mar 21 05:07:24 crc kubenswrapper[4775]: E0321 05:07:24.884755 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nb4h76h645hffh57dh684hc8h59chdh657h565hdhb9h5ch75h575h58h686h67h68bh679h65fh565h8bh5dbh9h5f8h596h555hd9h5b8h694q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tstsp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7bc489768c-rlbc8_openstack(44b58fbc-be53-4531-9dee-05ba398f25ef): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 05:07:24 crc kubenswrapper[4775]: E0321 05:07:24.887025 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7bc489768c-rlbc8" podUID="44b58fbc-be53-4531-9dee-05ba398f25ef" Mar 21 05:07:24 crc kubenswrapper[4775]: E0321 05:07:24.926565 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Mar 21 05:07:24 crc kubenswrapper[4775]: E0321 05:07:24.926722 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-764f7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-l8qqc_openstack(37bb7e34-ac47-44f6-b18f-ef4ed78eea6a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 05:07:24 crc kubenswrapper[4775]: E0321 05:07:24.928182 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-l8qqc" podUID="37bb7e34-ac47-44f6-b18f-ef4ed78eea6a" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.041379 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.042273 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-ovsdbserver-sb\") pod \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\" (UID: \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\") " Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.042310 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-dns-svc\") pod \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\" (UID: \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\") " Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.051918 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9fd475b9c-m48br" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.141606 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6b50afbd-31b8-40ff-bd7b-1ce5021e2837" (UID: "6b50afbd-31b8-40ff-bd7b-1ce5021e2837"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.143709 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltd4s\" (UniqueName: \"kubernetes.io/projected/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-kube-api-access-ltd4s\") pod \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\" (UID: \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\") " Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.143754 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-ovsdbserver-nb\") pod \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\" (UID: \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\") " Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.143785 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-config\") pod \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\" (UID: \"6b50afbd-31b8-40ff-bd7b-1ce5021e2837\") " Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.144588 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.145407 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6b50afbd-31b8-40ff-bd7b-1ce5021e2837" (UID: "6b50afbd-31b8-40ff-bd7b-1ce5021e2837"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.159901 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-kube-api-access-ltd4s" (OuterVolumeSpecName: "kube-api-access-ltd4s") pod "6b50afbd-31b8-40ff-bd7b-1ce5021e2837" (UID: "6b50afbd-31b8-40ff-bd7b-1ce5021e2837"). InnerVolumeSpecName "kube-api-access-ltd4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.198846 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-config" (OuterVolumeSpecName: "config") pod "6b50afbd-31b8-40ff-bd7b-1ce5021e2837" (UID: "6b50afbd-31b8-40ff-bd7b-1ce5021e2837"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.202631 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6b50afbd-31b8-40ff-bd7b-1ce5021e2837" (UID: "6b50afbd-31b8-40ff-bd7b-1ce5021e2837"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.245929 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f281a5f0-270c-4d95-93b1-ad265786b886-logs\") pod \"f281a5f0-270c-4d95-93b1-ad265786b886\" (UID: \"f281a5f0-270c-4d95-93b1-ad265786b886\") " Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.245998 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f281a5f0-270c-4d95-93b1-ad265786b886-scripts\") pod \"f281a5f0-270c-4d95-93b1-ad265786b886\" (UID: \"f281a5f0-270c-4d95-93b1-ad265786b886\") " Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.246093 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f281a5f0-270c-4d95-93b1-ad265786b886-horizon-secret-key\") pod \"f281a5f0-270c-4d95-93b1-ad265786b886\" (UID: \"f281a5f0-270c-4d95-93b1-ad265786b886\") " Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.246150 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjzgj\" (UniqueName: \"kubernetes.io/projected/f281a5f0-270c-4d95-93b1-ad265786b886-kube-api-access-jjzgj\") pod \"f281a5f0-270c-4d95-93b1-ad265786b886\" (UID: \"f281a5f0-270c-4d95-93b1-ad265786b886\") " Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.246252 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f281a5f0-270c-4d95-93b1-ad265786b886-config-data\") pod \"f281a5f0-270c-4d95-93b1-ad265786b886\" (UID: \"f281a5f0-270c-4d95-93b1-ad265786b886\") " Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.246336 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f281a5f0-270c-4d95-93b1-ad265786b886-logs" (OuterVolumeSpecName: "logs") pod "f281a5f0-270c-4d95-93b1-ad265786b886" (UID: "f281a5f0-270c-4d95-93b1-ad265786b886"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.246554 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f281a5f0-270c-4d95-93b1-ad265786b886-scripts" (OuterVolumeSpecName: "scripts") pod "f281a5f0-270c-4d95-93b1-ad265786b886" (UID: "f281a5f0-270c-4d95-93b1-ad265786b886"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.246902 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f281a5f0-270c-4d95-93b1-ad265786b886-config-data" (OuterVolumeSpecName: "config-data") pod "f281a5f0-270c-4d95-93b1-ad265786b886" (UID: "f281a5f0-270c-4d95-93b1-ad265786b886"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.246993 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f281a5f0-270c-4d95-93b1-ad265786b886-logs\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.247018 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f281a5f0-270c-4d95-93b1-ad265786b886-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.247033 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltd4s\" (UniqueName: \"kubernetes.io/projected/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-kube-api-access-ltd4s\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.247051 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.247065 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.247078 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b50afbd-31b8-40ff-bd7b-1ce5021e2837-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.249523 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f281a5f0-270c-4d95-93b1-ad265786b886-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "f281a5f0-270c-4d95-93b1-ad265786b886" (UID: "f281a5f0-270c-4d95-93b1-ad265786b886"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.250877 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f281a5f0-270c-4d95-93b1-ad265786b886-kube-api-access-jjzgj" (OuterVolumeSpecName: "kube-api-access-jjzgj") pod "f281a5f0-270c-4d95-93b1-ad265786b886" (UID: "f281a5f0-270c-4d95-93b1-ad265786b886"). InnerVolumeSpecName "kube-api-access-jjzgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.349344 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjzgj\" (UniqueName: \"kubernetes.io/projected/f281a5f0-270c-4d95-93b1-ad265786b886-kube-api-access-jjzgj\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.349382 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f281a5f0-270c-4d95-93b1-ad265786b886-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.349393 4775 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f281a5f0-270c-4d95-93b1-ad265786b886-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.381456 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6496ddbdd4-v5mc5"] Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.762361 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64fb567758-hd2ld"] Mar 21 05:07:25 crc kubenswrapper[4775]: W0321 05:07:25.763181 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09cf3763_0b41_4452_a247_d9a56f58b05d.slice/crio-7045bd948f55810ab12433655a5f2808bafab5e3e89d9d68cb00186c737c1405 WatchSource:0}: Error finding container 7045bd948f55810ab12433655a5f2808bafab5e3e89d9d68cb00186c737c1405: Status 404 returned error can't find the container with id 7045bd948f55810ab12433655a5f2808bafab5e3e89d9d68cb00186c737c1405 Mar 21 05:07:25 crc kubenswrapper[4775]: W0321 05:07:25.820535 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdf8b79b_6156_4aa5_a769_3a96408745c1.slice/crio-468779f2f3697fbd3ede621a392a75e5d47b2fbe74df31f239ef36748cd50133 WatchSource:0}: Error finding container 468779f2f3697fbd3ede621a392a75e5d47b2fbe74df31f239ef36748cd50133: Status 404 returned error can't find the container with id 468779f2f3697fbd3ede621a392a75e5d47b2fbe74df31f239ef36748cd50133 Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.824673 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jxmtf"] Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.830606 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" event={"ID":"6b50afbd-31b8-40ff-bd7b-1ce5021e2837","Type":"ContainerDied","Data":"2e772d1abd813f0f52a871bbd2aa2afc7b674d5369d7a3240dceb79e9ea9149f"} Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.830672 4775 scope.go:117] "RemoveContainer" containerID="6c721bd1cf19d6ad1fc2eb055e74ccf9afea13852ad17eeb02e87ea598c42465" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.830704 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.836738 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ac5c16c-56ac-4299-ae61-a8200986ce10","Type":"ContainerStarted","Data":"4e1e7c557d6fa58625eeced787347009a65ed5b8a3da8a6487f343ffff4e6c90"} Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.841262 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64fb567758-hd2ld" event={"ID":"09cf3763-0b41-4452-a247-d9a56f58b05d","Type":"ContainerStarted","Data":"7045bd948f55810ab12433655a5f2808bafab5e3e89d9d68cb00186c737c1405"} Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.842366 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6496ddbdd4-v5mc5" event={"ID":"fc6e433f-9e70-4b09-9780-403634bbe0dc","Type":"ContainerStarted","Data":"def71a90dc0cab6d68037e66af17622e205fe9b15e3446e1e10bd30605f49207"} Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.852717 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9fd475b9c-m48br" event={"ID":"f281a5f0-270c-4d95-93b1-ad265786b886","Type":"ContainerDied","Data":"3ed2947533338d4f3898375bd2e8d20bf0cdf56e90ab25944e2ce0199f079b4a"} Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.852898 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9fd475b9c-m48br" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.862959 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-b4m98"] Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.872163 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-b4m98"] Mar 21 05:07:25 crc kubenswrapper[4775]: E0321 05:07:25.929096 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-l8qqc" podUID="37bb7e34-ac47-44f6-b18f-ef4ed78eea6a" Mar 21 05:07:25 crc kubenswrapper[4775]: E0321 05:07:25.929417 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-9cmq2" podUID="1658991f-7a2b-4ce9-a240-a940385e0b8f" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.929487 4775 scope.go:117] "RemoveContainer" containerID="5d8a4462c4906d47c3863ff44ffe3802ba03cf186cd035ba4999f12c54eb2a8b" Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.964224 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-9fd475b9c-m48br"] Mar 21 05:07:25 crc kubenswrapper[4775]: I0321 05:07:25.973148 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-9fd475b9c-m48br"] Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.096823 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-b4m98" podUID="6b50afbd-31b8-40ff-bd7b-1ce5021e2837" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: i/o timeout" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.255646 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c4c5b5499-jnmhl" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.264330 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbdnp\" (UniqueName: \"kubernetes.io/projected/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-kube-api-access-rbdnp\") pod \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\" (UID: \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\") " Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.264778 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-logs\") pod \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\" (UID: \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\") " Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.264816 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-scripts\") pod \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\" (UID: \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\") " Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.264886 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-horizon-secret-key\") pod \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\" (UID: \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\") " Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.264943 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-config-data\") pod \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\" (UID: \"e7fbefc1-c1f6-472d-82b4-77c0a44b514f\") " Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.265199 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-logs" (OuterVolumeSpecName: "logs") pod "e7fbefc1-c1f6-472d-82b4-77c0a44b514f" (UID: "e7fbefc1-c1f6-472d-82b4-77c0a44b514f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.265735 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-logs\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.266808 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-config-data" (OuterVolumeSpecName: "config-data") pod "e7fbefc1-c1f6-472d-82b4-77c0a44b514f" (UID: "e7fbefc1-c1f6-472d-82b4-77c0a44b514f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.267489 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-scripts" (OuterVolumeSpecName: "scripts") pod "e7fbefc1-c1f6-472d-82b4-77c0a44b514f" (UID: "e7fbefc1-c1f6-472d-82b4-77c0a44b514f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.270353 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-kube-api-access-rbdnp" (OuterVolumeSpecName: "kube-api-access-rbdnp") pod "e7fbefc1-c1f6-472d-82b4-77c0a44b514f" (UID: "e7fbefc1-c1f6-472d-82b4-77c0a44b514f"). InnerVolumeSpecName "kube-api-access-rbdnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.276467 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "e7fbefc1-c1f6-472d-82b4-77c0a44b514f" (UID: "e7fbefc1-c1f6-472d-82b4-77c0a44b514f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.277529 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bc489768c-rlbc8" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.368542 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbdnp\" (UniqueName: \"kubernetes.io/projected/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-kube-api-access-rbdnp\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.368569 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.368581 4775 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.368591 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e7fbefc1-c1f6-472d-82b4-77c0a44b514f-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.470780 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tstsp\" (UniqueName: \"kubernetes.io/projected/44b58fbc-be53-4531-9dee-05ba398f25ef-kube-api-access-tstsp\") pod \"44b58fbc-be53-4531-9dee-05ba398f25ef\" (UID: \"44b58fbc-be53-4531-9dee-05ba398f25ef\") " Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.470860 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/44b58fbc-be53-4531-9dee-05ba398f25ef-scripts\") pod \"44b58fbc-be53-4531-9dee-05ba398f25ef\" (UID: \"44b58fbc-be53-4531-9dee-05ba398f25ef\") " Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.470893 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/44b58fbc-be53-4531-9dee-05ba398f25ef-config-data\") pod \"44b58fbc-be53-4531-9dee-05ba398f25ef\" (UID: \"44b58fbc-be53-4531-9dee-05ba398f25ef\") " Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.470940 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/44b58fbc-be53-4531-9dee-05ba398f25ef-horizon-secret-key\") pod \"44b58fbc-be53-4531-9dee-05ba398f25ef\" (UID: \"44b58fbc-be53-4531-9dee-05ba398f25ef\") " Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.470979 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44b58fbc-be53-4531-9dee-05ba398f25ef-logs\") pod \"44b58fbc-be53-4531-9dee-05ba398f25ef\" (UID: \"44b58fbc-be53-4531-9dee-05ba398f25ef\") " Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.471782 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44b58fbc-be53-4531-9dee-05ba398f25ef-logs" (OuterVolumeSpecName: "logs") pod "44b58fbc-be53-4531-9dee-05ba398f25ef" (UID: "44b58fbc-be53-4531-9dee-05ba398f25ef"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.471868 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44b58fbc-be53-4531-9dee-05ba398f25ef-scripts" (OuterVolumeSpecName: "scripts") pod "44b58fbc-be53-4531-9dee-05ba398f25ef" (UID: "44b58fbc-be53-4531-9dee-05ba398f25ef"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.472474 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44b58fbc-be53-4531-9dee-05ba398f25ef-config-data" (OuterVolumeSpecName: "config-data") pod "44b58fbc-be53-4531-9dee-05ba398f25ef" (UID: "44b58fbc-be53-4531-9dee-05ba398f25ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.475157 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44b58fbc-be53-4531-9dee-05ba398f25ef-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "44b58fbc-be53-4531-9dee-05ba398f25ef" (UID: "44b58fbc-be53-4531-9dee-05ba398f25ef"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.476251 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44b58fbc-be53-4531-9dee-05ba398f25ef-kube-api-access-tstsp" (OuterVolumeSpecName: "kube-api-access-tstsp") pod "44b58fbc-be53-4531-9dee-05ba398f25ef" (UID: "44b58fbc-be53-4531-9dee-05ba398f25ef"). InnerVolumeSpecName "kube-api-access-tstsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.573583 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tstsp\" (UniqueName: \"kubernetes.io/projected/44b58fbc-be53-4531-9dee-05ba398f25ef-kube-api-access-tstsp\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.573950 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/44b58fbc-be53-4531-9dee-05ba398f25ef-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.573966 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/44b58fbc-be53-4531-9dee-05ba398f25ef-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.573978 4775 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/44b58fbc-be53-4531-9dee-05ba398f25ef-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.573988 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44b58fbc-be53-4531-9dee-05ba398f25ef-logs\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.906035 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6496ddbdd4-v5mc5" event={"ID":"fc6e433f-9e70-4b09-9780-403634bbe0dc","Type":"ContainerStarted","Data":"8b8b8699c4dde7309a59e872dc53db96f0f40ebcc3885606c855e96185e66b67"} Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.906083 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6496ddbdd4-v5mc5" event={"ID":"fc6e433f-9e70-4b09-9780-403634bbe0dc","Type":"ContainerStarted","Data":"3f539614cc2828bc92642f654cb791c2699bf6b0605023f3c2de3117a0947231"} Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.910213 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c4c5b5499-jnmhl" event={"ID":"e7fbefc1-c1f6-472d-82b4-77c0a44b514f","Type":"ContainerDied","Data":"4d928750acf27a12d43a71655b2e716129e9bfb358dbdddb74f0ae57f062cbba"} Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.910216 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c4c5b5499-jnmhl" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.916465 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bc489768c-rlbc8" event={"ID":"44b58fbc-be53-4531-9dee-05ba398f25ef","Type":"ContainerDied","Data":"e60f7afbe17b7aa1bbfaf8e491a43aeb53519de12ebcef7dcaf29708ec522108"} Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.916556 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bc489768c-rlbc8" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.924025 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jxmtf" event={"ID":"cdf8b79b-6156-4aa5-a769-3a96408745c1","Type":"ContainerStarted","Data":"e72ab47d319ea50893e1c6b3a16586fa25056fa334f81da8dd2632328d94d338"} Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.924082 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jxmtf" event={"ID":"cdf8b79b-6156-4aa5-a769-3a96408745c1","Type":"ContainerStarted","Data":"468779f2f3697fbd3ede621a392a75e5d47b2fbe74df31f239ef36748cd50133"} Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.931198 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6496ddbdd4-v5mc5" podStartSLOduration=32.386875967 podStartE2EDuration="32.931180539s" podCreationTimestamp="2026-03-21 05:06:54 +0000 UTC" firstStartedPulling="2026-03-21 05:07:25.412782158 +0000 UTC m=+1198.389245782" lastFinishedPulling="2026-03-21 05:07:25.95708673 +0000 UTC m=+1198.933550354" observedRunningTime="2026-03-21 05:07:26.929691587 +0000 UTC m=+1199.906155231" watchObservedRunningTime="2026-03-21 05:07:26.931180539 +0000 UTC m=+1199.907644163" Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.931860 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64fb567758-hd2ld" event={"ID":"09cf3763-0b41-4452-a247-d9a56f58b05d","Type":"ContainerStarted","Data":"1da6670625f81794dc89cffb06ac6d008a9c7c8ed1b16d49c09b2332ff0e94fa"} Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.931912 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64fb567758-hd2ld" event={"ID":"09cf3763-0b41-4452-a247-d9a56f58b05d","Type":"ContainerStarted","Data":"4df1223c20d1958e562bffafa09f6a9587b632ec38a4f1dbf6a26792e8ce21fd"} Mar 21 05:07:26 crc kubenswrapper[4775]: I0321 05:07:26.966716 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-jxmtf" podStartSLOduration=4.966690433 podStartE2EDuration="4.966690433s" podCreationTimestamp="2026-03-21 05:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:07:26.951143993 +0000 UTC m=+1199.927607627" watchObservedRunningTime="2026-03-21 05:07:26.966690433 +0000 UTC m=+1199.943154057" Mar 21 05:07:27 crc kubenswrapper[4775]: I0321 05:07:27.005363 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7bc489768c-rlbc8"] Mar 21 05:07:27 crc kubenswrapper[4775]: I0321 05:07:27.023850 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7bc489768c-rlbc8"] Mar 21 05:07:27 crc kubenswrapper[4775]: I0321 05:07:27.053902 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6c4c5b5499-jnmhl"] Mar 21 05:07:27 crc kubenswrapper[4775]: I0321 05:07:27.061537 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6c4c5b5499-jnmhl"] Mar 21 05:07:27 crc kubenswrapper[4775]: I0321 05:07:27.066767 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-64fb567758-hd2ld" podStartSLOduration=32.615005353 podStartE2EDuration="33.06674658s" podCreationTimestamp="2026-03-21 05:06:54 +0000 UTC" firstStartedPulling="2026-03-21 05:07:25.765695511 +0000 UTC m=+1198.742159135" lastFinishedPulling="2026-03-21 05:07:26.217436738 +0000 UTC m=+1199.193900362" observedRunningTime="2026-03-21 05:07:27.033929313 +0000 UTC m=+1200.010392957" watchObservedRunningTime="2026-03-21 05:07:27.06674658 +0000 UTC m=+1200.043210204" Mar 21 05:07:27 crc kubenswrapper[4775]: I0321 05:07:27.676934 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44b58fbc-be53-4531-9dee-05ba398f25ef" path="/var/lib/kubelet/pods/44b58fbc-be53-4531-9dee-05ba398f25ef/volumes" Mar 21 05:07:27 crc kubenswrapper[4775]: I0321 05:07:27.677671 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b50afbd-31b8-40ff-bd7b-1ce5021e2837" path="/var/lib/kubelet/pods/6b50afbd-31b8-40ff-bd7b-1ce5021e2837/volumes" Mar 21 05:07:27 crc kubenswrapper[4775]: I0321 05:07:27.678338 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7fbefc1-c1f6-472d-82b4-77c0a44b514f" path="/var/lib/kubelet/pods/e7fbefc1-c1f6-472d-82b4-77c0a44b514f/volumes" Mar 21 05:07:27 crc kubenswrapper[4775]: I0321 05:07:27.678754 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f281a5f0-270c-4d95-93b1-ad265786b886" path="/var/lib/kubelet/pods/f281a5f0-270c-4d95-93b1-ad265786b886/volumes" Mar 21 05:07:31 crc kubenswrapper[4775]: I0321 05:07:31.984173 4775 generic.go:334] "Generic (PLEG): container finished" podID="716605f1-5111-4e7a-9591-18dfb5da1984" containerID="43def41a0cf1f318d43db6319e03e84d1a862d118a43d560e7c5e1d51f26dead" exitCode=0 Mar 21 05:07:31 crc kubenswrapper[4775]: I0321 05:07:31.984300 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-td76j" event={"ID":"716605f1-5111-4e7a-9591-18dfb5da1984","Type":"ContainerDied","Data":"43def41a0cf1f318d43db6319e03e84d1a862d118a43d560e7c5e1d51f26dead"} Mar 21 05:07:32 crc kubenswrapper[4775]: I0321 05:07:32.482768 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:07:32 crc kubenswrapper[4775]: I0321 05:07:32.482826 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:07:32 crc kubenswrapper[4775]: I0321 05:07:32.995482 4775 generic.go:334] "Generic (PLEG): container finished" podID="cdf8b79b-6156-4aa5-a769-3a96408745c1" containerID="e72ab47d319ea50893e1c6b3a16586fa25056fa334f81da8dd2632328d94d338" exitCode=0 Mar 21 05:07:32 crc kubenswrapper[4775]: I0321 05:07:32.995557 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jxmtf" event={"ID":"cdf8b79b-6156-4aa5-a769-3a96408745c1","Type":"ContainerDied","Data":"e72ab47d319ea50893e1c6b3a16586fa25056fa334f81da8dd2632328d94d338"} Mar 21 05:07:33 crc kubenswrapper[4775]: I0321 05:07:33.701187 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-td76j" Mar 21 05:07:33 crc kubenswrapper[4775]: I0321 05:07:33.817429 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/716605f1-5111-4e7a-9591-18dfb5da1984-db-sync-config-data\") pod \"716605f1-5111-4e7a-9591-18dfb5da1984\" (UID: \"716605f1-5111-4e7a-9591-18dfb5da1984\") " Mar 21 05:07:33 crc kubenswrapper[4775]: I0321 05:07:33.817592 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/716605f1-5111-4e7a-9591-18dfb5da1984-combined-ca-bundle\") pod \"716605f1-5111-4e7a-9591-18dfb5da1984\" (UID: \"716605f1-5111-4e7a-9591-18dfb5da1984\") " Mar 21 05:07:33 crc kubenswrapper[4775]: I0321 05:07:33.817647 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/716605f1-5111-4e7a-9591-18dfb5da1984-config-data\") pod \"716605f1-5111-4e7a-9591-18dfb5da1984\" (UID: \"716605f1-5111-4e7a-9591-18dfb5da1984\") " Mar 21 05:07:33 crc kubenswrapper[4775]: I0321 05:07:33.817723 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmggl\" (UniqueName: \"kubernetes.io/projected/716605f1-5111-4e7a-9591-18dfb5da1984-kube-api-access-zmggl\") pod \"716605f1-5111-4e7a-9591-18dfb5da1984\" (UID: \"716605f1-5111-4e7a-9591-18dfb5da1984\") " Mar 21 05:07:33 crc kubenswrapper[4775]: I0321 05:07:33.823256 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/716605f1-5111-4e7a-9591-18dfb5da1984-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "716605f1-5111-4e7a-9591-18dfb5da1984" (UID: "716605f1-5111-4e7a-9591-18dfb5da1984"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:33 crc kubenswrapper[4775]: I0321 05:07:33.823586 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/716605f1-5111-4e7a-9591-18dfb5da1984-kube-api-access-zmggl" (OuterVolumeSpecName: "kube-api-access-zmggl") pod "716605f1-5111-4e7a-9591-18dfb5da1984" (UID: "716605f1-5111-4e7a-9591-18dfb5da1984"). InnerVolumeSpecName "kube-api-access-zmggl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:07:33 crc kubenswrapper[4775]: I0321 05:07:33.840537 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/716605f1-5111-4e7a-9591-18dfb5da1984-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "716605f1-5111-4e7a-9591-18dfb5da1984" (UID: "716605f1-5111-4e7a-9591-18dfb5da1984"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:33 crc kubenswrapper[4775]: I0321 05:07:33.857338 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/716605f1-5111-4e7a-9591-18dfb5da1984-config-data" (OuterVolumeSpecName: "config-data") pod "716605f1-5111-4e7a-9591-18dfb5da1984" (UID: "716605f1-5111-4e7a-9591-18dfb5da1984"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:33 crc kubenswrapper[4775]: I0321 05:07:33.920300 4775 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/716605f1-5111-4e7a-9591-18dfb5da1984-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:33 crc kubenswrapper[4775]: I0321 05:07:33.920332 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/716605f1-5111-4e7a-9591-18dfb5da1984-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:33 crc kubenswrapper[4775]: I0321 05:07:33.920344 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/716605f1-5111-4e7a-9591-18dfb5da1984-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:33 crc kubenswrapper[4775]: I0321 05:07:33.920352 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmggl\" (UniqueName: \"kubernetes.io/projected/716605f1-5111-4e7a-9591-18dfb5da1984-kube-api-access-zmggl\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.022829 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ac5c16c-56ac-4299-ae61-a8200986ce10","Type":"ContainerStarted","Data":"3ffce781b6baa0a2f3ae718a0cd398633d05ff48719dba863adc5152b47fe849"} Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.025055 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-td76j" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.025190 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-td76j" event={"ID":"716605f1-5111-4e7a-9591-18dfb5da1984","Type":"ContainerDied","Data":"c9c92c1914b54290237327698e3251a34ba7a8a08da38fd6bd3a1901ce249b25"} Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.027404 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9c92c1914b54290237327698e3251a34ba7a8a08da38fd6bd3a1901ce249b25" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.380356 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-bvxtx"] Mar 21 05:07:34 crc kubenswrapper[4775]: E0321 05:07:34.380683 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b50afbd-31b8-40ff-bd7b-1ce5021e2837" containerName="init" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.380697 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b50afbd-31b8-40ff-bd7b-1ce5021e2837" containerName="init" Mar 21 05:07:34 crc kubenswrapper[4775]: E0321 05:07:34.380718 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="716605f1-5111-4e7a-9591-18dfb5da1984" containerName="glance-db-sync" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.380724 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="716605f1-5111-4e7a-9591-18dfb5da1984" containerName="glance-db-sync" Mar 21 05:07:34 crc kubenswrapper[4775]: E0321 05:07:34.380737 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b50afbd-31b8-40ff-bd7b-1ce5021e2837" containerName="dnsmasq-dns" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.380743 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b50afbd-31b8-40ff-bd7b-1ce5021e2837" containerName="dnsmasq-dns" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.380899 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="716605f1-5111-4e7a-9591-18dfb5da1984" containerName="glance-db-sync" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.380911 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b50afbd-31b8-40ff-bd7b-1ce5021e2837" containerName="dnsmasq-dns" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.381851 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.405383 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-bvxtx"] Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.490278 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.531061 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-bvxtx\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.531143 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-bvxtx\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.531286 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-bvxtx\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.531342 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-bvxtx\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.531367 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-config\") pod \"dnsmasq-dns-56df8fb6b7-bvxtx\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.531550 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vtjl\" (UniqueName: \"kubernetes.io/projected/f65ac8a3-4328-41af-8854-e8a2a9fce295-kube-api-access-6vtjl\") pod \"dnsmasq-dns-56df8fb6b7-bvxtx\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.632740 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-fernet-keys\") pod \"cdf8b79b-6156-4aa5-a769-3a96408745c1\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.632798 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dx7pf\" (UniqueName: \"kubernetes.io/projected/cdf8b79b-6156-4aa5-a769-3a96408745c1-kube-api-access-dx7pf\") pod \"cdf8b79b-6156-4aa5-a769-3a96408745c1\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.632834 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-scripts\") pod \"cdf8b79b-6156-4aa5-a769-3a96408745c1\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.632860 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-combined-ca-bundle\") pod \"cdf8b79b-6156-4aa5-a769-3a96408745c1\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.632944 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-config-data\") pod \"cdf8b79b-6156-4aa5-a769-3a96408745c1\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.633060 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-credential-keys\") pod \"cdf8b79b-6156-4aa5-a769-3a96408745c1\" (UID: \"cdf8b79b-6156-4aa5-a769-3a96408745c1\") " Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.633251 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-bvxtx\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.633322 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-bvxtx\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.633351 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-bvxtx\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.633368 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-config\") pod \"dnsmasq-dns-56df8fb6b7-bvxtx\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.633404 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vtjl\" (UniqueName: \"kubernetes.io/projected/f65ac8a3-4328-41af-8854-e8a2a9fce295-kube-api-access-6vtjl\") pod \"dnsmasq-dns-56df8fb6b7-bvxtx\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.633443 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-bvxtx\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.634605 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-config\") pod \"dnsmasq-dns-56df8fb6b7-bvxtx\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.634613 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-bvxtx\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.635579 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-bvxtx\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.635801 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-bvxtx\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.635988 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-bvxtx\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.641313 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-scripts" (OuterVolumeSpecName: "scripts") pod "cdf8b79b-6156-4aa5-a769-3a96408745c1" (UID: "cdf8b79b-6156-4aa5-a769-3a96408745c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.641370 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "cdf8b79b-6156-4aa5-a769-3a96408745c1" (UID: "cdf8b79b-6156-4aa5-a769-3a96408745c1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.644347 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdf8b79b-6156-4aa5-a769-3a96408745c1-kube-api-access-dx7pf" (OuterVolumeSpecName: "kube-api-access-dx7pf") pod "cdf8b79b-6156-4aa5-a769-3a96408745c1" (UID: "cdf8b79b-6156-4aa5-a769-3a96408745c1"). InnerVolumeSpecName "kube-api-access-dx7pf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.655224 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "cdf8b79b-6156-4aa5-a769-3a96408745c1" (UID: "cdf8b79b-6156-4aa5-a769-3a96408745c1"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.660840 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vtjl\" (UniqueName: \"kubernetes.io/projected/f65ac8a3-4328-41af-8854-e8a2a9fce295-kube-api-access-6vtjl\") pod \"dnsmasq-dns-56df8fb6b7-bvxtx\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.686174 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cdf8b79b-6156-4aa5-a769-3a96408745c1" (UID: "cdf8b79b-6156-4aa5-a769-3a96408745c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.712672 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-config-data" (OuterVolumeSpecName: "config-data") pod "cdf8b79b-6156-4aa5-a769-3a96408745c1" (UID: "cdf8b79b-6156-4aa5-a769-3a96408745c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.734651 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.734682 4775 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.734692 4775 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.734702 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dx7pf\" (UniqueName: \"kubernetes.io/projected/cdf8b79b-6156-4aa5-a769-3a96408745c1-kube-api-access-dx7pf\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.734711 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.734725 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf8b79b-6156-4aa5-a769-3a96408745c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.775419 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.979198 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:07:34 crc kubenswrapper[4775]: I0321 05:07:34.979494 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.032397 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.032444 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.033875 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jxmtf" event={"ID":"cdf8b79b-6156-4aa5-a769-3a96408745c1","Type":"ContainerDied","Data":"468779f2f3697fbd3ede621a392a75e5d47b2fbe74df31f239ef36748cd50133"} Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.033906 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="468779f2f3697fbd3ede621a392a75e5d47b2fbe74df31f239ef36748cd50133" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.033937 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jxmtf" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.199405 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-66c879cfdd-smnxp"] Mar 21 05:07:35 crc kubenswrapper[4775]: E0321 05:07:35.199832 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdf8b79b-6156-4aa5-a769-3a96408745c1" containerName="keystone-bootstrap" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.199855 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdf8b79b-6156-4aa5-a769-3a96408745c1" containerName="keystone-bootstrap" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.200094 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdf8b79b-6156-4aa5-a769-3a96408745c1" containerName="keystone-bootstrap" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.200776 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.203324 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.204015 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.204033 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.204054 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-nhwlw" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.204249 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.204554 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.219330 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-66c879cfdd-smnxp"] Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.230143 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-bvxtx"] Mar 21 05:07:35 crc kubenswrapper[4775]: W0321 05:07:35.237991 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf65ac8a3_4328_41af_8854_e8a2a9fce295.slice/crio-b2b7a3537eba6e1449d61c2042541c00aac4345e771c671fe8c458e056f81a53 WatchSource:0}: Error finding container b2b7a3537eba6e1449d61c2042541c00aac4345e771c671fe8c458e056f81a53: Status 404 returned error can't find the container with id b2b7a3537eba6e1449d61c2042541c00aac4345e771c671fe8c458e056f81a53 Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.279074 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.280843 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.286600 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.286790 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4x9cq" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.287057 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.288558 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.348180 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-credential-keys\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.348560 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-scripts\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.348596 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-public-tls-certs\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.348632 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-combined-ca-bundle\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.348654 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-config-data\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.348719 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-fernet-keys\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.348734 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjr9h\" (UniqueName: \"kubernetes.io/projected/9f965feb-5d82-4176-a14d-08a84c4ae794-kube-api-access-zjr9h\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.348777 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-internal-tls-certs\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.450619 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3627b61d-e88a-4523-8e6d-8a45a7e626c1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.450679 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-public-tls-certs\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.450786 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3627b61d-e88a-4523-8e6d-8a45a7e626c1-logs\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.450833 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3627b61d-e88a-4523-8e6d-8a45a7e626c1-config-data\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.450887 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f96cd\" (UniqueName: \"kubernetes.io/projected/3627b61d-e88a-4523-8e6d-8a45a7e626c1-kube-api-access-f96cd\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.450963 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-combined-ca-bundle\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.450994 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-config-data\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.451047 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-fernet-keys\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.451084 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjr9h\" (UniqueName: \"kubernetes.io/projected/9f965feb-5d82-4176-a14d-08a84c4ae794-kube-api-access-zjr9h\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.451111 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3627b61d-e88a-4523-8e6d-8a45a7e626c1-scripts\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.451235 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-internal-tls-certs\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.451323 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-credential-keys\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.451361 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.451485 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3627b61d-e88a-4523-8e6d-8a45a7e626c1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.451520 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-scripts\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.462389 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-config-data\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.462718 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-scripts\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.462809 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-internal-tls-certs\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.463313 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-public-tls-certs\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.463314 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-credential-keys\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.468291 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-fernet-keys\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.468618 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f965feb-5d82-4176-a14d-08a84c4ae794-combined-ca-bundle\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.477553 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjr9h\" (UniqueName: \"kubernetes.io/projected/9f965feb-5d82-4176-a14d-08a84c4ae794-kube-api-access-zjr9h\") pod \"keystone-66c879cfdd-smnxp\" (UID: \"9f965feb-5d82-4176-a14d-08a84c4ae794\") " pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.515013 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.553234 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.553324 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3627b61d-e88a-4523-8e6d-8a45a7e626c1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.553368 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3627b61d-e88a-4523-8e6d-8a45a7e626c1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.553392 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3627b61d-e88a-4523-8e6d-8a45a7e626c1-logs\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.553418 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3627b61d-e88a-4523-8e6d-8a45a7e626c1-config-data\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.553496 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f96cd\" (UniqueName: \"kubernetes.io/projected/3627b61d-e88a-4523-8e6d-8a45a7e626c1-kube-api-access-f96cd\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.553578 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.553593 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3627b61d-e88a-4523-8e6d-8a45a7e626c1-scripts\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.553882 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3627b61d-e88a-4523-8e6d-8a45a7e626c1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.555067 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3627b61d-e88a-4523-8e6d-8a45a7e626c1-logs\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.559390 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3627b61d-e88a-4523-8e6d-8a45a7e626c1-config-data\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.560713 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3627b61d-e88a-4523-8e6d-8a45a7e626c1-scripts\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.578785 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f96cd\" (UniqueName: \"kubernetes.io/projected/3627b61d-e88a-4523-8e6d-8a45a7e626c1-kube-api-access-f96cd\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.587013 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3627b61d-e88a-4523-8e6d-8a45a7e626c1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.592040 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.637647 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.640976 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.649660 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.649855 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.757181 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.757528 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.757561 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd8zc\" (UniqueName: \"kubernetes.io/projected/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-kube-api-access-hd8zc\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.757666 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.757715 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-logs\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.757754 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.757776 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.859388 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.859432 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.859451 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd8zc\" (UniqueName: \"kubernetes.io/projected/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-kube-api-access-hd8zc\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.859525 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.859558 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-logs\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.859585 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.859599 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.860290 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.860296 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-logs\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.861034 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.876271 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.876950 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.877078 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.884859 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd8zc\" (UniqueName: \"kubernetes.io/projected/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-kube-api-access-hd8zc\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.903459 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.908028 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-66c879cfdd-smnxp"] Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.949217 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:35 crc kubenswrapper[4775]: I0321 05:07:35.989134 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 05:07:36 crc kubenswrapper[4775]: I0321 05:07:36.053947 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dh4pj" event={"ID":"5cd9d81f-8c8a-46f9-9943-42dc0b638bef","Type":"ContainerStarted","Data":"1a60b462497324b6ca2e30820dbc175100c3b70b7e0a4ec6225ceb18cce50eaa"} Mar 21 05:07:36 crc kubenswrapper[4775]: I0321 05:07:36.059544 4775 generic.go:334] "Generic (PLEG): container finished" podID="f65ac8a3-4328-41af-8854-e8a2a9fce295" containerID="73687f66a6e199eea3343a248d8eb3f006e4d6f0bdcb64641db9e3b7a2d85758" exitCode=0 Mar 21 05:07:36 crc kubenswrapper[4775]: I0321 05:07:36.059632 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" event={"ID":"f65ac8a3-4328-41af-8854-e8a2a9fce295","Type":"ContainerDied","Data":"73687f66a6e199eea3343a248d8eb3f006e4d6f0bdcb64641db9e3b7a2d85758"} Mar 21 05:07:36 crc kubenswrapper[4775]: I0321 05:07:36.059681 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" event={"ID":"f65ac8a3-4328-41af-8854-e8a2a9fce295","Type":"ContainerStarted","Data":"b2b7a3537eba6e1449d61c2042541c00aac4345e771c671fe8c458e056f81a53"} Mar 21 05:07:36 crc kubenswrapper[4775]: I0321 05:07:36.066072 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-66c879cfdd-smnxp" event={"ID":"9f965feb-5d82-4176-a14d-08a84c4ae794","Type":"ContainerStarted","Data":"76ea5f3d3bd164fcd0cd09e31995b32bf21fe42671038c6ea3df00f04fb44076"} Mar 21 05:07:36 crc kubenswrapper[4775]: I0321 05:07:36.138817 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-dh4pj" podStartSLOduration=2.500218996 podStartE2EDuration="50.138799661s" podCreationTimestamp="2026-03-21 05:06:46 +0000 UTC" firstStartedPulling="2026-03-21 05:06:47.448641689 +0000 UTC m=+1160.425105313" lastFinishedPulling="2026-03-21 05:07:35.087222354 +0000 UTC m=+1208.063685978" observedRunningTime="2026-03-21 05:07:36.080612997 +0000 UTC m=+1209.057076621" watchObservedRunningTime="2026-03-21 05:07:36.138799661 +0000 UTC m=+1209.115263285" Mar 21 05:07:36 crc kubenswrapper[4775]: I0321 05:07:36.508004 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 05:07:36 crc kubenswrapper[4775]: I0321 05:07:36.598368 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 05:07:37 crc kubenswrapper[4775]: W0321 05:07:37.020509 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3627b61d_e88a_4523_8e6d_8a45a7e626c1.slice/crio-7327479ccc08ab30546cabbf16e6c59549cedbb1e41f1a39cfe62e456a6c97e2 WatchSource:0}: Error finding container 7327479ccc08ab30546cabbf16e6c59549cedbb1e41f1a39cfe62e456a6c97e2: Status 404 returned error can't find the container with id 7327479ccc08ab30546cabbf16e6c59549cedbb1e41f1a39cfe62e456a6c97e2 Mar 21 05:07:37 crc kubenswrapper[4775]: W0321 05:07:37.023370 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1c6e8e3_4fb9_472b_bbf1_2bdcd372bf34.slice/crio-84eb5f4cda9bda7b5808a7ebed1bb7d2eb4953a40585830ab493ee083b6e0bcf WatchSource:0}: Error finding container 84eb5f4cda9bda7b5808a7ebed1bb7d2eb4953a40585830ab493ee083b6e0bcf: Status 404 returned error can't find the container with id 84eb5f4cda9bda7b5808a7ebed1bb7d2eb4953a40585830ab493ee083b6e0bcf Mar 21 05:07:37 crc kubenswrapper[4775]: I0321 05:07:37.076235 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3627b61d-e88a-4523-8e6d-8a45a7e626c1","Type":"ContainerStarted","Data":"7327479ccc08ab30546cabbf16e6c59549cedbb1e41f1a39cfe62e456a6c97e2"} Mar 21 05:07:37 crc kubenswrapper[4775]: I0321 05:07:37.076982 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34","Type":"ContainerStarted","Data":"84eb5f4cda9bda7b5808a7ebed1bb7d2eb4953a40585830ab493ee083b6e0bcf"} Mar 21 05:07:38 crc kubenswrapper[4775]: I0321 05:07:38.017719 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 05:07:38 crc kubenswrapper[4775]: I0321 05:07:38.103109 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3627b61d-e88a-4523-8e6d-8a45a7e626c1","Type":"ContainerStarted","Data":"b0cd245fc2b0c70eaf76803ab7a80bbc622111a54e050582ef48868a191f3421"} Mar 21 05:07:38 crc kubenswrapper[4775]: I0321 05:07:38.106482 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-66c879cfdd-smnxp" event={"ID":"9f965feb-5d82-4176-a14d-08a84c4ae794","Type":"ContainerStarted","Data":"9e7c3bbccec13922e0bf4822134f7312831fdf48ec6372a3297f63367a30adc2"} Mar 21 05:07:38 crc kubenswrapper[4775]: I0321 05:07:38.109104 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:07:38 crc kubenswrapper[4775]: I0321 05:07:38.115652 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34","Type":"ContainerStarted","Data":"32f21565b6fc51221eac38731ca38e1d9609cbf9ce961697af6d2d9ccdf3f74f"} Mar 21 05:07:38 crc kubenswrapper[4775]: I0321 05:07:38.123907 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 05:07:38 crc kubenswrapper[4775]: I0321 05:07:38.138037 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-66c879cfdd-smnxp" podStartSLOduration=3.138017901 podStartE2EDuration="3.138017901s" podCreationTimestamp="2026-03-21 05:07:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:07:38.13129087 +0000 UTC m=+1211.107754504" watchObservedRunningTime="2026-03-21 05:07:38.138017901 +0000 UTC m=+1211.114481525" Mar 21 05:07:38 crc kubenswrapper[4775]: I0321 05:07:38.155160 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" event={"ID":"f65ac8a3-4328-41af-8854-e8a2a9fce295","Type":"ContainerStarted","Data":"18c26c2a6ec8c71c7881c713a6d2fb65eaa2258d47fd0f272a51e57881466532"} Mar 21 05:07:38 crc kubenswrapper[4775]: I0321 05:07:38.155795 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:38 crc kubenswrapper[4775]: I0321 05:07:38.202079 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" podStartSLOduration=4.20204844 podStartE2EDuration="4.20204844s" podCreationTimestamp="2026-03-21 05:07:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:07:38.172423943 +0000 UTC m=+1211.148887577" watchObservedRunningTime="2026-03-21 05:07:38.20204844 +0000 UTC m=+1211.178512064" Mar 21 05:07:39 crc kubenswrapper[4775]: I0321 05:07:39.165350 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3627b61d-e88a-4523-8e6d-8a45a7e626c1","Type":"ContainerStarted","Data":"a8cd516d29617897d19857166b7b9b4d04fb227e81a99f9300ea48c6eed5c3c4"} Mar 21 05:07:39 crc kubenswrapper[4775]: I0321 05:07:39.165677 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3627b61d-e88a-4523-8e6d-8a45a7e626c1" containerName="glance-log" containerID="cri-o://b0cd245fc2b0c70eaf76803ab7a80bbc622111a54e050582ef48868a191f3421" gracePeriod=30 Mar 21 05:07:39 crc kubenswrapper[4775]: I0321 05:07:39.165956 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3627b61d-e88a-4523-8e6d-8a45a7e626c1" containerName="glance-httpd" containerID="cri-o://a8cd516d29617897d19857166b7b9b4d04fb227e81a99f9300ea48c6eed5c3c4" gracePeriod=30 Mar 21 05:07:39 crc kubenswrapper[4775]: I0321 05:07:39.168785 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34","Type":"ContainerStarted","Data":"1d80db6f5c73575575e757ede386b167af7f87a0b631372385a77aa72299bb8b"} Mar 21 05:07:39 crc kubenswrapper[4775]: I0321 05:07:39.168947 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34" containerName="glance-log" containerID="cri-o://32f21565b6fc51221eac38731ca38e1d9609cbf9ce961697af6d2d9ccdf3f74f" gracePeriod=30 Mar 21 05:07:39 crc kubenswrapper[4775]: I0321 05:07:39.169009 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34" containerName="glance-httpd" containerID="cri-o://1d80db6f5c73575575e757ede386b167af7f87a0b631372385a77aa72299bb8b" gracePeriod=30 Mar 21 05:07:39 crc kubenswrapper[4775]: I0321 05:07:39.192489 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.192474039 podStartE2EDuration="5.192474039s" podCreationTimestamp="2026-03-21 05:07:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:07:39.191037689 +0000 UTC m=+1212.167501313" watchObservedRunningTime="2026-03-21 05:07:39.192474039 +0000 UTC m=+1212.168937663" Mar 21 05:07:39 crc kubenswrapper[4775]: I0321 05:07:39.218229 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.218207216 podStartE2EDuration="5.218207216s" podCreationTimestamp="2026-03-21 05:07:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:07:39.214315737 +0000 UTC m=+1212.190779371" watchObservedRunningTime="2026-03-21 05:07:39.218207216 +0000 UTC m=+1212.194670840" Mar 21 05:07:40 crc kubenswrapper[4775]: I0321 05:07:40.199494 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-l8qqc" event={"ID":"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a","Type":"ContainerStarted","Data":"84b76b926432d603dd6d10be15e6cfc9e20ac65787b9fb2a30c3bdbd38a96e09"} Mar 21 05:07:40 crc kubenswrapper[4775]: I0321 05:07:40.205178 4775 generic.go:334] "Generic (PLEG): container finished" podID="3627b61d-e88a-4523-8e6d-8a45a7e626c1" containerID="a8cd516d29617897d19857166b7b9b4d04fb227e81a99f9300ea48c6eed5c3c4" exitCode=0 Mar 21 05:07:40 crc kubenswrapper[4775]: I0321 05:07:40.205215 4775 generic.go:334] "Generic (PLEG): container finished" podID="3627b61d-e88a-4523-8e6d-8a45a7e626c1" containerID="b0cd245fc2b0c70eaf76803ab7a80bbc622111a54e050582ef48868a191f3421" exitCode=143 Mar 21 05:07:40 crc kubenswrapper[4775]: I0321 05:07:40.205250 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3627b61d-e88a-4523-8e6d-8a45a7e626c1","Type":"ContainerDied","Data":"a8cd516d29617897d19857166b7b9b4d04fb227e81a99f9300ea48c6eed5c3c4"} Mar 21 05:07:40 crc kubenswrapper[4775]: I0321 05:07:40.205293 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3627b61d-e88a-4523-8e6d-8a45a7e626c1","Type":"ContainerDied","Data":"b0cd245fc2b0c70eaf76803ab7a80bbc622111a54e050582ef48868a191f3421"} Mar 21 05:07:40 crc kubenswrapper[4775]: I0321 05:07:40.208782 4775 generic.go:334] "Generic (PLEG): container finished" podID="9531f5f9-8f77-4882-b779-4210b1de81de" containerID="f19a3eb592d811463bd26fefd306c16b1ec6187913716e5585ce5b02489e2c0c" exitCode=0 Mar 21 05:07:40 crc kubenswrapper[4775]: I0321 05:07:40.208984 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wvjlb" event={"ID":"9531f5f9-8f77-4882-b779-4210b1de81de","Type":"ContainerDied","Data":"f19a3eb592d811463bd26fefd306c16b1ec6187913716e5585ce5b02489e2c0c"} Mar 21 05:07:40 crc kubenswrapper[4775]: I0321 05:07:40.217821 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-l8qqc" podStartSLOduration=4.012702732 podStartE2EDuration="55.217799636s" podCreationTimestamp="2026-03-21 05:06:45 +0000 UTC" firstStartedPulling="2026-03-21 05:06:47.033579901 +0000 UTC m=+1160.010043525" lastFinishedPulling="2026-03-21 05:07:38.238676805 +0000 UTC m=+1211.215140429" observedRunningTime="2026-03-21 05:07:40.216304664 +0000 UTC m=+1213.192768298" watchObservedRunningTime="2026-03-21 05:07:40.217799636 +0000 UTC m=+1213.194263260" Mar 21 05:07:40 crc kubenswrapper[4775]: I0321 05:07:40.218723 4775 generic.go:334] "Generic (PLEG): container finished" podID="e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34" containerID="1d80db6f5c73575575e757ede386b167af7f87a0b631372385a77aa72299bb8b" exitCode=0 Mar 21 05:07:40 crc kubenswrapper[4775]: I0321 05:07:40.218755 4775 generic.go:334] "Generic (PLEG): container finished" podID="e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34" containerID="32f21565b6fc51221eac38731ca38e1d9609cbf9ce961697af6d2d9ccdf3f74f" exitCode=143 Mar 21 05:07:40 crc kubenswrapper[4775]: I0321 05:07:40.222575 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34","Type":"ContainerDied","Data":"1d80db6f5c73575575e757ede386b167af7f87a0b631372385a77aa72299bb8b"} Mar 21 05:07:40 crc kubenswrapper[4775]: I0321 05:07:40.222630 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34","Type":"ContainerDied","Data":"32f21565b6fc51221eac38731ca38e1d9609cbf9ce961697af6d2d9ccdf3f74f"} Mar 21 05:07:41 crc kubenswrapper[4775]: I0321 05:07:41.233276 4775 generic.go:334] "Generic (PLEG): container finished" podID="5cd9d81f-8c8a-46f9-9943-42dc0b638bef" containerID="1a60b462497324b6ca2e30820dbc175100c3b70b7e0a4ec6225ceb18cce50eaa" exitCode=0 Mar 21 05:07:41 crc kubenswrapper[4775]: I0321 05:07:41.233661 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dh4pj" event={"ID":"5cd9d81f-8c8a-46f9-9943-42dc0b638bef","Type":"ContainerDied","Data":"1a60b462497324b6ca2e30820dbc175100c3b70b7e0a4ec6225ceb18cce50eaa"} Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.779997 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.895944 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3627b61d-e88a-4523-8e6d-8a45a7e626c1-logs\") pod \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.896003 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3627b61d-e88a-4523-8e6d-8a45a7e626c1-config-data\") pod \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.896027 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3627b61d-e88a-4523-8e6d-8a45a7e626c1-combined-ca-bundle\") pod \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.896256 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3627b61d-e88a-4523-8e6d-8a45a7e626c1-httpd-run\") pod \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.896335 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3627b61d-e88a-4523-8e6d-8a45a7e626c1-scripts\") pod \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.896396 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f96cd\" (UniqueName: \"kubernetes.io/projected/3627b61d-e88a-4523-8e6d-8a45a7e626c1-kube-api-access-f96cd\") pod \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.896437 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\" (UID: \"3627b61d-e88a-4523-8e6d-8a45a7e626c1\") " Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.898162 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3627b61d-e88a-4523-8e6d-8a45a7e626c1-logs" (OuterVolumeSpecName: "logs") pod "3627b61d-e88a-4523-8e6d-8a45a7e626c1" (UID: "3627b61d-e88a-4523-8e6d-8a45a7e626c1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.898400 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3627b61d-e88a-4523-8e6d-8a45a7e626c1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3627b61d-e88a-4523-8e6d-8a45a7e626c1" (UID: "3627b61d-e88a-4523-8e6d-8a45a7e626c1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.906355 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3627b61d-e88a-4523-8e6d-8a45a7e626c1-scripts" (OuterVolumeSpecName: "scripts") pod "3627b61d-e88a-4523-8e6d-8a45a7e626c1" (UID: "3627b61d-e88a-4523-8e6d-8a45a7e626c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.913580 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3627b61d-e88a-4523-8e6d-8a45a7e626c1-kube-api-access-f96cd" (OuterVolumeSpecName: "kube-api-access-f96cd") pod "3627b61d-e88a-4523-8e6d-8a45a7e626c1" (UID: "3627b61d-e88a-4523-8e6d-8a45a7e626c1"). InnerVolumeSpecName "kube-api-access-f96cd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.918447 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "3627b61d-e88a-4523-8e6d-8a45a7e626c1" (UID: "3627b61d-e88a-4523-8e6d-8a45a7e626c1"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.937224 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3627b61d-e88a-4523-8e6d-8a45a7e626c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3627b61d-e88a-4523-8e6d-8a45a7e626c1" (UID: "3627b61d-e88a-4523-8e6d-8a45a7e626c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.959284 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3627b61d-e88a-4523-8e6d-8a45a7e626c1-config-data" (OuterVolumeSpecName: "config-data") pod "3627b61d-e88a-4523-8e6d-8a45a7e626c1" (UID: "3627b61d-e88a-4523-8e6d-8a45a7e626c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.998111 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3627b61d-e88a-4523-8e6d-8a45a7e626c1-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.998154 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3627b61d-e88a-4523-8e6d-8a45a7e626c1-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.998163 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f96cd\" (UniqueName: \"kubernetes.io/projected/3627b61d-e88a-4523-8e6d-8a45a7e626c1-kube-api-access-f96cd\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.998193 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.998203 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3627b61d-e88a-4523-8e6d-8a45a7e626c1-logs\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.998211 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3627b61d-e88a-4523-8e6d-8a45a7e626c1-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:42 crc kubenswrapper[4775]: I0321 05:07:42.998219 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3627b61d-e88a-4523-8e6d-8a45a7e626c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.032340 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.099678 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.252015 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3627b61d-e88a-4523-8e6d-8a45a7e626c1","Type":"ContainerDied","Data":"7327479ccc08ab30546cabbf16e6c59549cedbb1e41f1a39cfe62e456a6c97e2"} Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.252079 4775 scope.go:117] "RemoveContainer" containerID="a8cd516d29617897d19857166b7b9b4d04fb227e81a99f9300ea48c6eed5c3c4" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.252237 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.291078 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.301661 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.325074 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 05:07:43 crc kubenswrapper[4775]: E0321 05:07:43.325471 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3627b61d-e88a-4523-8e6d-8a45a7e626c1" containerName="glance-httpd" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.325489 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3627b61d-e88a-4523-8e6d-8a45a7e626c1" containerName="glance-httpd" Mar 21 05:07:43 crc kubenswrapper[4775]: E0321 05:07:43.325518 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3627b61d-e88a-4523-8e6d-8a45a7e626c1" containerName="glance-log" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.325524 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3627b61d-e88a-4523-8e6d-8a45a7e626c1" containerName="glance-log" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.325691 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3627b61d-e88a-4523-8e6d-8a45a7e626c1" containerName="glance-httpd" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.325721 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3627b61d-e88a-4523-8e6d-8a45a7e626c1" containerName="glance-log" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.326645 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.331392 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.332526 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.333235 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.407997 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-scripts\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.408109 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/47e577d7-e389-4135-b4fb-979bd627eaa9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.408294 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.408339 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4spw\" (UniqueName: \"kubernetes.io/projected/47e577d7-e389-4135-b4fb-979bd627eaa9-kube-api-access-d4spw\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.408400 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.408446 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.408504 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-config-data\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.408565 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47e577d7-e389-4135-b4fb-979bd627eaa9-logs\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.510469 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-scripts\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.510527 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/47e577d7-e389-4135-b4fb-979bd627eaa9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.510617 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.510671 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4spw\" (UniqueName: \"kubernetes.io/projected/47e577d7-e389-4135-b4fb-979bd627eaa9-kube-api-access-d4spw\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.510713 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.510743 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.510779 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-config-data\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.510815 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47e577d7-e389-4135-b4fb-979bd627eaa9-logs\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.511021 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/47e577d7-e389-4135-b4fb-979bd627eaa9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.511321 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.511500 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47e577d7-e389-4135-b4fb-979bd627eaa9-logs\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.517174 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-scripts\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.517465 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.517528 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-config-data\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.520209 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.538958 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4spw\" (UniqueName: \"kubernetes.io/projected/47e577d7-e389-4135-b4fb-979bd627eaa9-kube-api-access-d4spw\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.542647 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.647043 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 05:07:43 crc kubenswrapper[4775]: I0321 05:07:43.673202 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3627b61d-e88a-4523-8e6d-8a45a7e626c1" path="/var/lib/kubelet/pods/3627b61d-e88a-4523-8e6d-8a45a7e626c1/volumes" Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.513201 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.629231 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-scripts\") pod \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.629337 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hd8zc\" (UniqueName: \"kubernetes.io/projected/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-kube-api-access-hd8zc\") pod \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.629371 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-config-data\") pod \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.629405 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.629432 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-combined-ca-bundle\") pod \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.629564 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-httpd-run\") pod \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.629602 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-logs\") pod \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\" (UID: \"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34\") " Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.630381 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34" (UID: "e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.630415 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-logs" (OuterVolumeSpecName: "logs") pod "e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34" (UID: "e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.633918 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-scripts" (OuterVolumeSpecName: "scripts") pod "e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34" (UID: "e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.635071 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34" (UID: "e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.635717 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-kube-api-access-hd8zc" (OuterVolumeSpecName: "kube-api-access-hd8zc") pod "e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34" (UID: "e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34"). InnerVolumeSpecName "kube-api-access-hd8zc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.667629 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34" (UID: "e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.685829 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-config-data" (OuterVolumeSpecName: "config-data") pod "e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34" (UID: "e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.732965 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.733018 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hd8zc\" (UniqueName: \"kubernetes.io/projected/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-kube-api-access-hd8zc\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.733033 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.733063 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.733075 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.733087 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.733375 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34-logs\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.752846 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.777783 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.836051 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.843915 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4nfhc"] Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.844339 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" podUID="aaad9349-8b1e-4f07-b3c9-36bc4781a386" containerName="dnsmasq-dns" containerID="cri-o://f2fafd7848327783c5d640051878058b230f83637eec544e85f17df14ee71f99" gracePeriod=10 Mar 21 05:07:44 crc kubenswrapper[4775]: I0321 05:07:44.987445 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-64fb567758-hd2ld" podUID="09cf3763-0b41-4452-a247-d9a56f58b05d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.035398 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6496ddbdd4-v5mc5" podUID="fc6e433f-9e70-4b09-9780-403634bbe0dc" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.273649 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34","Type":"ContainerDied","Data":"84eb5f4cda9bda7b5808a7ebed1bb7d2eb4953a40585830ab493ee083b6e0bcf"} Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.273665 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.277724 4775 generic.go:334] "Generic (PLEG): container finished" podID="aaad9349-8b1e-4f07-b3c9-36bc4781a386" containerID="f2fafd7848327783c5d640051878058b230f83637eec544e85f17df14ee71f99" exitCode=0 Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.277764 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" event={"ID":"aaad9349-8b1e-4f07-b3c9-36bc4781a386","Type":"ContainerDied","Data":"f2fafd7848327783c5d640051878058b230f83637eec544e85f17df14ee71f99"} Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.340677 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.354484 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.369835 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 05:07:45 crc kubenswrapper[4775]: E0321 05:07:45.370418 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34" containerName="glance-httpd" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.370439 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34" containerName="glance-httpd" Mar 21 05:07:45 crc kubenswrapper[4775]: E0321 05:07:45.370454 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34" containerName="glance-log" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.370500 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34" containerName="glance-log" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.370738 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34" containerName="glance-httpd" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.370760 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34" containerName="glance-log" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.371937 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.374932 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.374981 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.378550 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.447717 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.447764 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/71255c4f-1e47-4e35-845f-876fff5fd6d4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.447845 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71255c4f-1e47-4e35-845f-876fff5fd6d4-logs\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.447864 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz2pk\" (UniqueName: \"kubernetes.io/projected/71255c4f-1e47-4e35-845f-876fff5fd6d4-kube-api-access-pz2pk\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.447894 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.447924 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.447953 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.447971 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.549695 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.549742 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/71255c4f-1e47-4e35-845f-876fff5fd6d4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.549827 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71255c4f-1e47-4e35-845f-876fff5fd6d4-logs\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.549845 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz2pk\" (UniqueName: \"kubernetes.io/projected/71255c4f-1e47-4e35-845f-876fff5fd6d4-kube-api-access-pz2pk\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.549873 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.549903 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.549928 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.549946 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.550128 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.550876 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/71255c4f-1e47-4e35-845f-876fff5fd6d4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.553021 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71255c4f-1e47-4e35-845f-876fff5fd6d4-logs\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.554936 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.555547 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.555752 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.559732 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.572978 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz2pk\" (UniqueName: \"kubernetes.io/projected/71255c4f-1e47-4e35-845f-876fff5fd6d4-kube-api-access-pz2pk\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.578354 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.672776 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34" path="/var/lib/kubelet/pods/e1c6e8e3-4fb9-472b-bbf1-2bdcd372bf34/volumes" Mar 21 05:07:45 crc kubenswrapper[4775]: I0321 05:07:45.733011 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 05:07:46 crc kubenswrapper[4775]: I0321 05:07:46.287210 4775 generic.go:334] "Generic (PLEG): container finished" podID="37bb7e34-ac47-44f6-b18f-ef4ed78eea6a" containerID="84b76b926432d603dd6d10be15e6cfc9e20ac65787b9fb2a30c3bdbd38a96e09" exitCode=0 Mar 21 05:07:46 crc kubenswrapper[4775]: I0321 05:07:46.287293 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-l8qqc" event={"ID":"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a","Type":"ContainerDied","Data":"84b76b926432d603dd6d10be15e6cfc9e20ac65787b9fb2a30c3bdbd38a96e09"} Mar 21 05:07:46 crc kubenswrapper[4775]: I0321 05:07:46.700618 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" podUID="aaad9349-8b1e-4f07-b3c9-36bc4781a386" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.145:5353: connect: connection refused" Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.604964 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wvjlb" Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.610664 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dh4pj" Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.688495 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-scripts\") pod \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\" (UID: \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\") " Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.688589 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9531f5f9-8f77-4882-b779-4210b1de81de-combined-ca-bundle\") pod \"9531f5f9-8f77-4882-b779-4210b1de81de\" (UID: \"9531f5f9-8f77-4882-b779-4210b1de81de\") " Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.688750 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-logs\") pod \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\" (UID: \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\") " Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.688895 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-config-data\") pod \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\" (UID: \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\") " Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.688932 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9531f5f9-8f77-4882-b779-4210b1de81de-config\") pod \"9531f5f9-8f77-4882-b779-4210b1de81de\" (UID: \"9531f5f9-8f77-4882-b779-4210b1de81de\") " Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.688973 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf22x\" (UniqueName: \"kubernetes.io/projected/9531f5f9-8f77-4882-b779-4210b1de81de-kube-api-access-gf22x\") pod \"9531f5f9-8f77-4882-b779-4210b1de81de\" (UID: \"9531f5f9-8f77-4882-b779-4210b1de81de\") " Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.689030 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-combined-ca-bundle\") pod \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\" (UID: \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\") " Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.689071 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgm22\" (UniqueName: \"kubernetes.io/projected/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-kube-api-access-sgm22\") pod \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\" (UID: \"5cd9d81f-8c8a-46f9-9943-42dc0b638bef\") " Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.689565 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-logs" (OuterVolumeSpecName: "logs") pod "5cd9d81f-8c8a-46f9-9943-42dc0b638bef" (UID: "5cd9d81f-8c8a-46f9-9943-42dc0b638bef"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.689871 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-logs\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.696193 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-kube-api-access-sgm22" (OuterVolumeSpecName: "kube-api-access-sgm22") pod "5cd9d81f-8c8a-46f9-9943-42dc0b638bef" (UID: "5cd9d81f-8c8a-46f9-9943-42dc0b638bef"). InnerVolumeSpecName "kube-api-access-sgm22". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.697099 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9531f5f9-8f77-4882-b779-4210b1de81de-kube-api-access-gf22x" (OuterVolumeSpecName: "kube-api-access-gf22x") pod "9531f5f9-8f77-4882-b779-4210b1de81de" (UID: "9531f5f9-8f77-4882-b779-4210b1de81de"). InnerVolumeSpecName "kube-api-access-gf22x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.698504 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-scripts" (OuterVolumeSpecName: "scripts") pod "5cd9d81f-8c8a-46f9-9943-42dc0b638bef" (UID: "5cd9d81f-8c8a-46f9-9943-42dc0b638bef"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.715683 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-config-data" (OuterVolumeSpecName: "config-data") pod "5cd9d81f-8c8a-46f9-9943-42dc0b638bef" (UID: "5cd9d81f-8c8a-46f9-9943-42dc0b638bef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.716553 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5cd9d81f-8c8a-46f9-9943-42dc0b638bef" (UID: "5cd9d81f-8c8a-46f9-9943-42dc0b638bef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.718266 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9531f5f9-8f77-4882-b779-4210b1de81de-config" (OuterVolumeSpecName: "config") pod "9531f5f9-8f77-4882-b779-4210b1de81de" (UID: "9531f5f9-8f77-4882-b779-4210b1de81de"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.734091 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9531f5f9-8f77-4882-b779-4210b1de81de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9531f5f9-8f77-4882-b779-4210b1de81de" (UID: "9531f5f9-8f77-4882-b779-4210b1de81de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.792130 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.792168 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/9531f5f9-8f77-4882-b779-4210b1de81de-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.792181 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf22x\" (UniqueName: \"kubernetes.io/projected/9531f5f9-8f77-4882-b779-4210b1de81de-kube-api-access-gf22x\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.792195 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.792209 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgm22\" (UniqueName: \"kubernetes.io/projected/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-kube-api-access-sgm22\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.792220 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cd9d81f-8c8a-46f9-9943-42dc0b638bef-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:47 crc kubenswrapper[4775]: I0321 05:07:47.792231 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9531f5f9-8f77-4882-b779-4210b1de81de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.305377 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wvjlb" event={"ID":"9531f5f9-8f77-4882-b779-4210b1de81de","Type":"ContainerDied","Data":"a0dba2a88554c5f818ac14151546ad519e642afe620cd4a8580aa68fd7107a3c"} Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.305754 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0dba2a88554c5f818ac14151546ad519e642afe620cd4a8580aa68fd7107a3c" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.305483 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wvjlb" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.307911 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dh4pj" event={"ID":"5cd9d81f-8c8a-46f9-9943-42dc0b638bef","Type":"ContainerDied","Data":"f10dda8724855df58c36f76f7e138a9aa8f4ab4f82a8c7d9fc8bb963f6bad44c"} Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.307958 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f10dda8724855df58c36f76f7e138a9aa8f4ab4f82a8c7d9fc8bb963f6bad44c" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.308007 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dh4pj" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.620895 4775 scope.go:117] "RemoveContainer" containerID="b0cd245fc2b0c70eaf76803ab7a80bbc622111a54e050582ef48868a191f3421" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.827178 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-58948d8bb4-rcw89"] Mar 21 05:07:48 crc kubenswrapper[4775]: E0321 05:07:48.827840 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9531f5f9-8f77-4882-b779-4210b1de81de" containerName="neutron-db-sync" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.827853 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="9531f5f9-8f77-4882-b779-4210b1de81de" containerName="neutron-db-sync" Mar 21 05:07:48 crc kubenswrapper[4775]: E0321 05:07:48.827874 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cd9d81f-8c8a-46f9-9943-42dc0b638bef" containerName="placement-db-sync" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.827880 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cd9d81f-8c8a-46f9-9943-42dc0b638bef" containerName="placement-db-sync" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.828035 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="9531f5f9-8f77-4882-b779-4210b1de81de" containerName="neutron-db-sync" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.828062 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cd9d81f-8c8a-46f9-9943-42dc0b638bef" containerName="placement-db-sync" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.828991 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.831360 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.831468 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.831576 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.831744 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-8q2g8" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.831887 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.846873 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-58948d8bb4-rcw89"] Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.859475 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-lk5mk"] Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.860789 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.913698 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-lk5mk"] Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.914732 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-lk5mk\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.914791 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/279dff90-9d39-418a-b5e7-00333a376d16-internal-tls-certs\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.914823 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2r94w\" (UniqueName: \"kubernetes.io/projected/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-kube-api-access-2r94w\") pod \"dnsmasq-dns-6b7b667979-lk5mk\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.914847 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjz8j\" (UniqueName: \"kubernetes.io/projected/279dff90-9d39-418a-b5e7-00333a376d16-kube-api-access-sjz8j\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.914877 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/279dff90-9d39-418a-b5e7-00333a376d16-scripts\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.914903 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/279dff90-9d39-418a-b5e7-00333a376d16-config-data\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.914946 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-config\") pod \"dnsmasq-dns-6b7b667979-lk5mk\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.915018 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/279dff90-9d39-418a-b5e7-00333a376d16-combined-ca-bundle\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.915045 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/279dff90-9d39-418a-b5e7-00333a376d16-public-tls-certs\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.915073 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/279dff90-9d39-418a-b5e7-00333a376d16-logs\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.915095 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-dns-svc\") pod \"dnsmasq-dns-6b7b667979-lk5mk\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.915141 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-lk5mk\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.915168 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-lk5mk\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:48 crc kubenswrapper[4775]: I0321 05:07:48.979828 4775 scope.go:117] "RemoveContainer" containerID="1d80db6f5c73575575e757ede386b167af7f87a0b631372385a77aa72299bb8b" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.017644 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-lk5mk\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.028470 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-lk5mk\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.028584 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/279dff90-9d39-418a-b5e7-00333a376d16-internal-tls-certs\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.028656 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2r94w\" (UniqueName: \"kubernetes.io/projected/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-kube-api-access-2r94w\") pod \"dnsmasq-dns-6b7b667979-lk5mk\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.028703 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjz8j\" (UniqueName: \"kubernetes.io/projected/279dff90-9d39-418a-b5e7-00333a376d16-kube-api-access-sjz8j\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.028759 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/279dff90-9d39-418a-b5e7-00333a376d16-scripts\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.022869 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-lk5mk\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.022896 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-68ccf5bf68-lf5dz"] Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.030409 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-lk5mk\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.028814 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/279dff90-9d39-418a-b5e7-00333a376d16-config-data\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.030682 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-config\") pod \"dnsmasq-dns-6b7b667979-lk5mk\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.030957 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/279dff90-9d39-418a-b5e7-00333a376d16-combined-ca-bundle\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.031192 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/279dff90-9d39-418a-b5e7-00333a376d16-public-tls-certs\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.032223 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/279dff90-9d39-418a-b5e7-00333a376d16-logs\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.032362 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-dns-svc\") pod \"dnsmasq-dns-6b7b667979-lk5mk\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.032508 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-lk5mk\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.033730 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/279dff90-9d39-418a-b5e7-00333a376d16-logs\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.036813 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-dns-svc\") pod \"dnsmasq-dns-6b7b667979-lk5mk\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.039170 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-lk5mk\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.039585 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-config\") pod \"dnsmasq-dns-6b7b667979-lk5mk\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.043959 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.047174 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-68ccf5bf68-lf5dz"] Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.054081 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-pms25" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.054419 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.054435 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.054650 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.059025 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2r94w\" (UniqueName: \"kubernetes.io/projected/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-kube-api-access-2r94w\") pod \"dnsmasq-dns-6b7b667979-lk5mk\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.059598 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.060306 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/279dff90-9d39-418a-b5e7-00333a376d16-config-data\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.063749 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjz8j\" (UniqueName: \"kubernetes.io/projected/279dff90-9d39-418a-b5e7-00333a376d16-kube-api-access-sjz8j\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.063834 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/279dff90-9d39-418a-b5e7-00333a376d16-public-tls-certs\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.091697 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/279dff90-9d39-418a-b5e7-00333a376d16-scripts\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.097694 4775 scope.go:117] "RemoveContainer" containerID="32f21565b6fc51221eac38731ca38e1d9609cbf9ce961697af6d2d9ccdf3f74f" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.098399 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/279dff90-9d39-418a-b5e7-00333a376d16-internal-tls-certs\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.100706 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/279dff90-9d39-418a-b5e7-00333a376d16-combined-ca-bundle\") pod \"placement-58948d8bb4-rcw89\" (UID: \"279dff90-9d39-418a-b5e7-00333a376d16\") " pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.136455 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-764f7\" (UniqueName: \"kubernetes.io/projected/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-kube-api-access-764f7\") pod \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.136529 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-combined-ca-bundle\") pod \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.136565 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-scripts\") pod \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.136597 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-config-data\") pod \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.136756 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-etc-machine-id\") pod \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.136863 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-db-sync-config-data\") pod \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\" (UID: \"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a\") " Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.137105 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-ovndb-tls-certs\") pod \"neutron-68ccf5bf68-lf5dz\" (UID: \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\") " pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.137160 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-httpd-config\") pod \"neutron-68ccf5bf68-lf5dz\" (UID: \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\") " pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.137192 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-combined-ca-bundle\") pod \"neutron-68ccf5bf68-lf5dz\" (UID: \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\") " pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.137310 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-config\") pod \"neutron-68ccf5bf68-lf5dz\" (UID: \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\") " pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.137359 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv44z\" (UniqueName: \"kubernetes.io/projected/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-kube-api-access-bv44z\") pod \"neutron-68ccf5bf68-lf5dz\" (UID: \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\") " pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.137451 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "37bb7e34-ac47-44f6-b18f-ef4ed78eea6a" (UID: "37bb7e34-ac47-44f6-b18f-ef4ed78eea6a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.143806 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-kube-api-access-764f7" (OuterVolumeSpecName: "kube-api-access-764f7") pod "37bb7e34-ac47-44f6-b18f-ef4ed78eea6a" (UID: "37bb7e34-ac47-44f6-b18f-ef4ed78eea6a"). InnerVolumeSpecName "kube-api-access-764f7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.146414 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-scripts" (OuterVolumeSpecName: "scripts") pod "37bb7e34-ac47-44f6-b18f-ef4ed78eea6a" (UID: "37bb7e34-ac47-44f6-b18f-ef4ed78eea6a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.146659 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.150030 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.153291 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "37bb7e34-ac47-44f6-b18f-ef4ed78eea6a" (UID: "37bb7e34-ac47-44f6-b18f-ef4ed78eea6a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.166645 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:49 crc kubenswrapper[4775]: E0321 05:07:49.202858 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="8ac5c16c-56ac-4299-ae61-a8200986ce10" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.216340 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37bb7e34-ac47-44f6-b18f-ef4ed78eea6a" (UID: "37bb7e34-ac47-44f6-b18f-ef4ed78eea6a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.239745 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-ovsdbserver-sb\") pod \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.239852 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-config\") pod \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.239886 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-dns-svc\") pod \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.239920 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzx7f\" (UniqueName: \"kubernetes.io/projected/aaad9349-8b1e-4f07-b3c9-36bc4781a386-kube-api-access-lzx7f\") pod \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.239985 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-ovsdbserver-nb\") pod \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.240049 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-dns-swift-storage-0\") pod \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\" (UID: \"aaad9349-8b1e-4f07-b3c9-36bc4781a386\") " Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.240755 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-config\") pod \"neutron-68ccf5bf68-lf5dz\" (UID: \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\") " pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.240891 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv44z\" (UniqueName: \"kubernetes.io/projected/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-kube-api-access-bv44z\") pod \"neutron-68ccf5bf68-lf5dz\" (UID: \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\") " pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.240956 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-ovndb-tls-certs\") pod \"neutron-68ccf5bf68-lf5dz\" (UID: \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\") " pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.241003 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-httpd-config\") pod \"neutron-68ccf5bf68-lf5dz\" (UID: \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\") " pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.241041 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-combined-ca-bundle\") pod \"neutron-68ccf5bf68-lf5dz\" (UID: \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\") " pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.241240 4775 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.241258 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-764f7\" (UniqueName: \"kubernetes.io/projected/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-kube-api-access-764f7\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.241271 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.241282 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.241294 4775 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.244311 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-config-data" (OuterVolumeSpecName: "config-data") pod "37bb7e34-ac47-44f6-b18f-ef4ed78eea6a" (UID: "37bb7e34-ac47-44f6-b18f-ef4ed78eea6a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.247550 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-ovndb-tls-certs\") pod \"neutron-68ccf5bf68-lf5dz\" (UID: \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\") " pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.250784 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaad9349-8b1e-4f07-b3c9-36bc4781a386-kube-api-access-lzx7f" (OuterVolumeSpecName: "kube-api-access-lzx7f") pod "aaad9349-8b1e-4f07-b3c9-36bc4781a386" (UID: "aaad9349-8b1e-4f07-b3c9-36bc4781a386"). InnerVolumeSpecName "kube-api-access-lzx7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.258694 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-httpd-config\") pod \"neutron-68ccf5bf68-lf5dz\" (UID: \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\") " pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.258996 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-combined-ca-bundle\") pod \"neutron-68ccf5bf68-lf5dz\" (UID: \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\") " pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.267748 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv44z\" (UniqueName: \"kubernetes.io/projected/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-kube-api-access-bv44z\") pod \"neutron-68ccf5bf68-lf5dz\" (UID: \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\") " pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.269432 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-config\") pod \"neutron-68ccf5bf68-lf5dz\" (UID: \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\") " pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.305734 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aaad9349-8b1e-4f07-b3c9-36bc4781a386" (UID: "aaad9349-8b1e-4f07-b3c9-36bc4781a386"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.309337 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-config" (OuterVolumeSpecName: "config") pod "aaad9349-8b1e-4f07-b3c9-36bc4781a386" (UID: "aaad9349-8b1e-4f07-b3c9-36bc4781a386"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.316150 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aaad9349-8b1e-4f07-b3c9-36bc4781a386" (UID: "aaad9349-8b1e-4f07-b3c9-36bc4781a386"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.316555 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aaad9349-8b1e-4f07-b3c9-36bc4781a386" (UID: "aaad9349-8b1e-4f07-b3c9-36bc4781a386"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.327433 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9cmq2" event={"ID":"1658991f-7a2b-4ce9-a240-a940385e0b8f","Type":"ContainerStarted","Data":"796b087daa0094c51c318e24d71b74b2beb873149bcfc37bff7fc5e4bd3c8ed7"} Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.333038 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "aaad9349-8b1e-4f07-b3c9-36bc4781a386" (UID: "aaad9349-8b1e-4f07-b3c9-36bc4781a386"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.333292 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ac5c16c-56ac-4299-ae61-a8200986ce10","Type":"ContainerStarted","Data":"507f523f41fa20fba933e438c92908136440aa69787cc4ade1ab3f5729729a60"} Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.333534 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.333493 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8ac5c16c-56ac-4299-ae61-a8200986ce10" containerName="sg-core" containerID="cri-o://3ffce781b6baa0a2f3ae718a0cd398633d05ff48719dba863adc5152b47fe849" gracePeriod=30 Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.333387 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8ac5c16c-56ac-4299-ae61-a8200986ce10" containerName="ceilometer-notification-agent" containerID="cri-o://4e1e7c557d6fa58625eeced787347009a65ed5b8a3da8a6487f343ffff4e6c90" gracePeriod=30 Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.333506 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8ac5c16c-56ac-4299-ae61-a8200986ce10" containerName="proxy-httpd" containerID="cri-o://507f523f41fa20fba933e438c92908136440aa69787cc4ade1ab3f5729729a60" gracePeriod=30 Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.339645 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" event={"ID":"aaad9349-8b1e-4f07-b3c9-36bc4781a386","Type":"ContainerDied","Data":"545d0a4dfbecffb0f9c29e4cf099e382302c9c5682c08fe8a7fbe8e6c0247a35"} Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.339740 4775 scope.go:117] "RemoveContainer" containerID="f2fafd7848327783c5d640051878058b230f83637eec544e85f17df14ee71f99" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.339912 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-4nfhc" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.353918 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.354499 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.354561 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.354615 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.354667 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzx7f\" (UniqueName: \"kubernetes.io/projected/aaad9349-8b1e-4f07-b3c9-36bc4781a386-kube-api-access-lzx7f\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.354736 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.354795 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aaad9349-8b1e-4f07-b3c9-36bc4781a386-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.355008 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-l8qqc" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.355341 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-l8qqc" event={"ID":"37bb7e34-ac47-44f6-b18f-ef4ed78eea6a","Type":"ContainerDied","Data":"45cbfcc3ba69b1d165e2a6acc84f20765ed5bc3bad1c098418f5af9089eb0bf1"} Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.355418 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45cbfcc3ba69b1d165e2a6acc84f20765ed5bc3bad1c098418f5af9089eb0bf1" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.359773 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-9cmq2" podStartSLOduration=2.187875791 podStartE2EDuration="1m3.359752203s" podCreationTimestamp="2026-03-21 05:06:46 +0000 UTC" firstStartedPulling="2026-03-21 05:06:47.494573548 +0000 UTC m=+1160.471037182" lastFinishedPulling="2026-03-21 05:07:48.66644995 +0000 UTC m=+1221.642913594" observedRunningTime="2026-03-21 05:07:49.346064436 +0000 UTC m=+1222.322528060" watchObservedRunningTime="2026-03-21 05:07:49.359752203 +0000 UTC m=+1222.336215827" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.375226 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.385981 4775 scope.go:117] "RemoveContainer" containerID="4bd42a0432eaefd26033f0ffdb594981243afbef4a05205089435812a0be336b" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.434798 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4nfhc"] Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.444698 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4nfhc"] Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.485521 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.512759 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.625196 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-lk5mk"] Mar 21 05:07:49 crc kubenswrapper[4775]: W0321 05:07:49.674977 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5a4d8e0_d0d0_4b04_9f21_ddf0d94c0e8f.slice/crio-fbe39bc0c49faa915d2c4f50c757fc038107aae3b2ab6a7a97e1e90b02b1d1a4 WatchSource:0}: Error finding container fbe39bc0c49faa915d2c4f50c757fc038107aae3b2ab6a7a97e1e90b02b1d1a4: Status 404 returned error can't find the container with id fbe39bc0c49faa915d2c4f50c757fc038107aae3b2ab6a7a97e1e90b02b1d1a4 Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.704593 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aaad9349-8b1e-4f07-b3c9-36bc4781a386" path="/var/lib/kubelet/pods/aaad9349-8b1e-4f07-b3c9-36bc4781a386/volumes" Mar 21 05:07:49 crc kubenswrapper[4775]: I0321 05:07:49.710504 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-58948d8bb4-rcw89"] Mar 21 05:07:49 crc kubenswrapper[4775]: W0321 05:07:49.719897 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod279dff90_9d39_418a_b5e7_00333a376d16.slice/crio-7e253dae2491da166ef4e31ce862a09f50e642f143ea394133c2a881166deda6 WatchSource:0}: Error finding container 7e253dae2491da166ef4e31ce862a09f50e642f143ea394133c2a881166deda6: Status 404 returned error can't find the container with id 7e253dae2491da166ef4e31ce862a09f50e642f143ea394133c2a881166deda6 Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.153618 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-68ccf5bf68-lf5dz"] Mar 21 05:07:50 crc kubenswrapper[4775]: W0321 05:07:50.154881 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6dc5e1ad_635a_478c_8fac_ac1fcdb2bad4.slice/crio-785cf1f0611121bc184a4bc0f34377c346c490853673595b9763600cad5d4149 WatchSource:0}: Error finding container 785cf1f0611121bc184a4bc0f34377c346c490853673595b9763600cad5d4149: Status 404 returned error can't find the container with id 785cf1f0611121bc184a4bc0f34377c346c490853673595b9763600cad5d4149 Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.314269 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 21 05:07:50 crc kubenswrapper[4775]: E0321 05:07:50.314733 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaad9349-8b1e-4f07-b3c9-36bc4781a386" containerName="dnsmasq-dns" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.314759 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaad9349-8b1e-4f07-b3c9-36bc4781a386" containerName="dnsmasq-dns" Mar 21 05:07:50 crc kubenswrapper[4775]: E0321 05:07:50.314776 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaad9349-8b1e-4f07-b3c9-36bc4781a386" containerName="init" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.314784 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaad9349-8b1e-4f07-b3c9-36bc4781a386" containerName="init" Mar 21 05:07:50 crc kubenswrapper[4775]: E0321 05:07:50.314813 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37bb7e34-ac47-44f6-b18f-ef4ed78eea6a" containerName="cinder-db-sync" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.314823 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="37bb7e34-ac47-44f6-b18f-ef4ed78eea6a" containerName="cinder-db-sync" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.315037 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="37bb7e34-ac47-44f6-b18f-ef4ed78eea6a" containerName="cinder-db-sync" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.315075 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaad9349-8b1e-4f07-b3c9-36bc4781a386" containerName="dnsmasq-dns" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.316189 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.324606 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-nm9jl" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.324858 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.325104 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.325262 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.329558 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.375038 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f765864-79ee-415c-99be-55894a20c087-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " pod="openstack/cinder-scheduler-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.375311 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " pod="openstack/cinder-scheduler-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.375339 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-scripts\") pod \"cinder-scheduler-0\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " pod="openstack/cinder-scheduler-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.375438 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-config-data\") pod \"cinder-scheduler-0\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " pod="openstack/cinder-scheduler-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.375457 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " pod="openstack/cinder-scheduler-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.375485 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krwll\" (UniqueName: \"kubernetes.io/projected/3f765864-79ee-415c-99be-55894a20c087-kube-api-access-krwll\") pod \"cinder-scheduler-0\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " pod="openstack/cinder-scheduler-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.417859 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-lk5mk"] Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.421432 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"71255c4f-1e47-4e35-845f-876fff5fd6d4","Type":"ContainerStarted","Data":"b0bf1b828206099564568160f6163284ee71379761dd593fcfafe3139a339268"} Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.421641 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"71255c4f-1e47-4e35-845f-876fff5fd6d4","Type":"ContainerStarted","Data":"5d037f60a7382103283fb7dc977007a3668afac9d065f14a09fa2c15aa21b79e"} Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.463418 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47e577d7-e389-4135-b4fb-979bd627eaa9","Type":"ContainerStarted","Data":"ea4d39e040f38b33347813ecac429d1646c7747cc95596bc69a68249c6432f4a"} Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.477492 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f765864-79ee-415c-99be-55894a20c087-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " pod="openstack/cinder-scheduler-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.478024 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " pod="openstack/cinder-scheduler-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.478110 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-scripts\") pod \"cinder-scheduler-0\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " pod="openstack/cinder-scheduler-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.478324 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-config-data\") pod \"cinder-scheduler-0\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " pod="openstack/cinder-scheduler-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.478402 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " pod="openstack/cinder-scheduler-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.478486 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krwll\" (UniqueName: \"kubernetes.io/projected/3f765864-79ee-415c-99be-55894a20c087-kube-api-access-krwll\") pod \"cinder-scheduler-0\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " pod="openstack/cinder-scheduler-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.478863 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f765864-79ee-415c-99be-55894a20c087-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " pod="openstack/cinder-scheduler-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.482174 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-774db89647-9pnrf"] Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.483905 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.489884 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " pod="openstack/cinder-scheduler-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.496096 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-config-data\") pod \"cinder-scheduler-0\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " pod="openstack/cinder-scheduler-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.500752 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-scripts\") pod \"cinder-scheduler-0\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " pod="openstack/cinder-scheduler-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.509205 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " pod="openstack/cinder-scheduler-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.521205 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-774db89647-9pnrf"] Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.524937 4775 generic.go:334] "Generic (PLEG): container finished" podID="8ac5c16c-56ac-4299-ae61-a8200986ce10" containerID="3ffce781b6baa0a2f3ae718a0cd398633d05ff48719dba863adc5152b47fe849" exitCode=2 Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.525009 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ac5c16c-56ac-4299-ae61-a8200986ce10","Type":"ContainerDied","Data":"3ffce781b6baa0a2f3ae718a0cd398633d05ff48719dba863adc5152b47fe849"} Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.526614 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krwll\" (UniqueName: \"kubernetes.io/projected/3f765864-79ee-415c-99be-55894a20c087-kube-api-access-krwll\") pod \"cinder-scheduler-0\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " pod="openstack/cinder-scheduler-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.557476 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68ccf5bf68-lf5dz" event={"ID":"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4","Type":"ContainerStarted","Data":"f14a2e346ace37be895dcc94dd25c65e41716e7e754b07d2e1e3277674aeac90"} Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.557516 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68ccf5bf68-lf5dz" event={"ID":"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4","Type":"ContainerStarted","Data":"785cf1f0611121bc184a4bc0f34377c346c490853673595b9763600cad5d4149"} Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.558868 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-58948d8bb4-rcw89" event={"ID":"279dff90-9d39-418a-b5e7-00333a376d16","Type":"ContainerStarted","Data":"e3c653ccf6756c923a2ae6b032f17712a859f9775c3f8a1f6d64239a81f63d2e"} Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.558889 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-58948d8bb4-rcw89" event={"ID":"279dff90-9d39-418a-b5e7-00333a376d16","Type":"ContainerStarted","Data":"548faf90447b1a3a4e1b70a575543b3ccd563821c6c627bcd784e95cf7f5a15f"} Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.558897 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-58948d8bb4-rcw89" event={"ID":"279dff90-9d39-418a-b5e7-00333a376d16","Type":"ContainerStarted","Data":"7e253dae2491da166ef4e31ce862a09f50e642f143ea394133c2a881166deda6"} Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.559199 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.559241 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.653700 4775 generic.go:334] "Generic (PLEG): container finished" podID="f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f" containerID="180859970b041857dafe2a96bb332f8d1442cf9ef9839e6bbd7b85783c3dc75d" exitCode=0 Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.653885 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" event={"ID":"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f","Type":"ContainerDied","Data":"180859970b041857dafe2a96bb332f8d1442cf9ef9839e6bbd7b85783c3dc75d"} Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.653912 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" event={"ID":"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f","Type":"ContainerStarted","Data":"fbe39bc0c49faa915d2c4f50c757fc038107aae3b2ab6a7a97e1e90b02b1d1a4"} Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.654851 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-dns-swift-storage-0\") pod \"dnsmasq-dns-774db89647-9pnrf\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.655028 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-config\") pod \"dnsmasq-dns-774db89647-9pnrf\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.657898 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-dns-svc\") pod \"dnsmasq-dns-774db89647-9pnrf\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.660773 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.665232 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4hqb\" (UniqueName: \"kubernetes.io/projected/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-kube-api-access-c4hqb\") pod \"dnsmasq-dns-774db89647-9pnrf\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.665339 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-ovsdbserver-nb\") pod \"dnsmasq-dns-774db89647-9pnrf\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.665461 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-ovsdbserver-sb\") pod \"dnsmasq-dns-774db89647-9pnrf\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.737069 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-58948d8bb4-rcw89" podStartSLOduration=2.7370444860000003 podStartE2EDuration="2.737044486s" podCreationTimestamp="2026-03-21 05:07:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:07:50.644285214 +0000 UTC m=+1223.620748838" watchObservedRunningTime="2026-03-21 05:07:50.737044486 +0000 UTC m=+1223.713508110" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.771920 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-config\") pod \"dnsmasq-dns-774db89647-9pnrf\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.772325 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-dns-svc\") pod \"dnsmasq-dns-774db89647-9pnrf\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.772351 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4hqb\" (UniqueName: \"kubernetes.io/projected/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-kube-api-access-c4hqb\") pod \"dnsmasq-dns-774db89647-9pnrf\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.772405 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-ovsdbserver-nb\") pod \"dnsmasq-dns-774db89647-9pnrf\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.772461 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-ovsdbserver-sb\") pod \"dnsmasq-dns-774db89647-9pnrf\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.772501 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-dns-swift-storage-0\") pod \"dnsmasq-dns-774db89647-9pnrf\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.795711 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-dns-swift-storage-0\") pod \"dnsmasq-dns-774db89647-9pnrf\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.805933 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-dns-svc\") pod \"dnsmasq-dns-774db89647-9pnrf\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.806870 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-config\") pod \"dnsmasq-dns-774db89647-9pnrf\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.807519 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-ovsdbserver-nb\") pod \"dnsmasq-dns-774db89647-9pnrf\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.808582 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-ovsdbserver-sb\") pod \"dnsmasq-dns-774db89647-9pnrf\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.838084 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4hqb\" (UniqueName: \"kubernetes.io/projected/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-kube-api-access-c4hqb\") pod \"dnsmasq-dns-774db89647-9pnrf\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.858587 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.860166 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.872290 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.889098 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.906146 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.979197 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-config-data\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.979576 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb53eb05-afa8-4039-805c-477b25ed3075-logs\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.979715 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.979917 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-config-data-custom\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.980068 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb53eb05-afa8-4039-805c-477b25ed3075-etc-machine-id\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.980156 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctknn\" (UniqueName: \"kubernetes.io/projected/cb53eb05-afa8-4039-805c-477b25ed3075-kube-api-access-ctknn\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:50 crc kubenswrapper[4775]: I0321 05:07:50.980250 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-scripts\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.082451 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-config-data\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.082530 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb53eb05-afa8-4039-805c-477b25ed3075-logs\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.082554 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.082590 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-config-data-custom\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.082661 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb53eb05-afa8-4039-805c-477b25ed3075-etc-machine-id\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.082691 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctknn\" (UniqueName: \"kubernetes.io/projected/cb53eb05-afa8-4039-805c-477b25ed3075-kube-api-access-ctknn\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.082723 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-scripts\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.083974 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb53eb05-afa8-4039-805c-477b25ed3075-logs\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.084591 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb53eb05-afa8-4039-805c-477b25ed3075-etc-machine-id\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.088137 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-config-data\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.088833 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-scripts\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.091602 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.092027 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-config-data-custom\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.105039 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctknn\" (UniqueName: \"kubernetes.io/projected/cb53eb05-afa8-4039-805c-477b25ed3075-kube-api-access-ctknn\") pod \"cinder-api-0\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " pod="openstack/cinder-api-0" Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.219885 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 21 05:07:51 crc kubenswrapper[4775]: E0321 05:07:51.308160 4775 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Mar 21 05:07:51 crc kubenswrapper[4775]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Mar 21 05:07:51 crc kubenswrapper[4775]: > podSandboxID="fbe39bc0c49faa915d2c4f50c757fc038107aae3b2ab6a7a97e1e90b02b1d1a4" Mar 21 05:07:51 crc kubenswrapper[4775]: E0321 05:07:51.308368 4775 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 21 05:07:51 crc kubenswrapper[4775]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n548h78h598h67bh67chf7h67dhch88h5bh596h546h5b9h5b7h57hd9h58h5dbh674hfch548h549h99hbch5c9h546h687h687h648hd4h8bh98q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-swift-storage-0,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-swift-storage-0,SubPath:dns-swift-storage-0,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2r94w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6b7b667979-lk5mk_openstack(f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Mar 21 05:07:51 crc kubenswrapper[4775]: > logger="UnhandledError" Mar 21 05:07:51 crc kubenswrapper[4775]: E0321 05:07:51.309518 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" podUID="f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f" Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.339932 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 21 05:07:51 crc kubenswrapper[4775]: W0321 05:07:51.394548 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f765864_79ee_415c_99be_55894a20c087.slice/crio-85e53e5965bf199e0c4123a3954f4e5d437587f60519d3c96ac82b5f50bb3679 WatchSource:0}: Error finding container 85e53e5965bf199e0c4123a3954f4e5d437587f60519d3c96ac82b5f50bb3679: Status 404 returned error can't find the container with id 85e53e5965bf199e0c4123a3954f4e5d437587f60519d3c96ac82b5f50bb3679 Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.492893 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-774db89647-9pnrf"] Mar 21 05:07:51 crc kubenswrapper[4775]: W0321 05:07:51.509198 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b34a259_4754_41ea_b3b7_cd1ba5fb53c6.slice/crio-a73be0fd90826150202a28a361d434d06f770ab26ca7c2e2a7f4cd5e030488a3 WatchSource:0}: Error finding container a73be0fd90826150202a28a361d434d06f770ab26ca7c2e2a7f4cd5e030488a3: Status 404 returned error can't find the container with id a73be0fd90826150202a28a361d434d06f770ab26ca7c2e2a7f4cd5e030488a3 Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.682368 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47e577d7-e389-4135-b4fb-979bd627eaa9","Type":"ContainerStarted","Data":"0173c3dcb2f1ee362b645abaeead4a322f137b6409341a3b746ad5f19cffafb5"} Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.682966 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47e577d7-e389-4135-b4fb-979bd627eaa9","Type":"ContainerStarted","Data":"29a18bd4308c28984ec8d5e5e490324db58708c1d07372b24ee7783e1995157d"} Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.685833 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68ccf5bf68-lf5dz" event={"ID":"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4","Type":"ContainerStarted","Data":"06c961a967b1c07bf826487fd3c050eb92ff3106f515b9136df5283020977b92"} Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.685973 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.691705 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"71255c4f-1e47-4e35-845f-876fff5fd6d4","Type":"ContainerStarted","Data":"9f447f1b0e6151bf34a0eec832692e899e0a001912e2cf06411e6494eecfb405"} Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.694437 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-774db89647-9pnrf" event={"ID":"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6","Type":"ContainerStarted","Data":"a73be0fd90826150202a28a361d434d06f770ab26ca7c2e2a7f4cd5e030488a3"} Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.705561 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3f765864-79ee-415c-99be-55894a20c087","Type":"ContainerStarted","Data":"85e53e5965bf199e0c4123a3954f4e5d437587f60519d3c96ac82b5f50bb3679"} Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.709706 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.711625 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.711609828 podStartE2EDuration="8.711609828s" podCreationTimestamp="2026-03-21 05:07:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:07:51.695339198 +0000 UTC m=+1224.671802822" watchObservedRunningTime="2026-03-21 05:07:51.711609828 +0000 UTC m=+1224.688073452" Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.756066 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.7560453339999995 podStartE2EDuration="6.756045334s" podCreationTimestamp="2026-03-21 05:07:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:07:51.71945889 +0000 UTC m=+1224.695922524" watchObservedRunningTime="2026-03-21 05:07:51.756045334 +0000 UTC m=+1224.732508968" Mar 21 05:07:51 crc kubenswrapper[4775]: I0321 05:07:51.761441 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-68ccf5bf68-lf5dz" podStartSLOduration=3.761423136 podStartE2EDuration="3.761423136s" podCreationTimestamp="2026-03-21 05:07:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:07:51.746180845 +0000 UTC m=+1224.722644479" watchObservedRunningTime="2026-03-21 05:07:51.761423136 +0000 UTC m=+1224.737886760" Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.099485 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.206171 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-config\") pod \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.206237 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2r94w\" (UniqueName: \"kubernetes.io/projected/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-kube-api-access-2r94w\") pod \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.206300 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-dns-swift-storage-0\") pod \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.206319 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-ovsdbserver-sb\") pod \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.206461 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-dns-svc\") pod \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.206503 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-ovsdbserver-nb\") pod \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\" (UID: \"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f\") " Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.214324 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-kube-api-access-2r94w" (OuterVolumeSpecName: "kube-api-access-2r94w") pod "f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f" (UID: "f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f"). InnerVolumeSpecName "kube-api-access-2r94w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.261031 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f" (UID: "f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.264791 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-config" (OuterVolumeSpecName: "config") pod "f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f" (UID: "f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.266568 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f" (UID: "f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.266701 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f" (UID: "f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.284089 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f" (UID: "f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.307988 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.308014 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.308024 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.308034 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2r94w\" (UniqueName: \"kubernetes.io/projected/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-kube-api-access-2r94w\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.308044 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.308052 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.725281 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" event={"ID":"f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f","Type":"ContainerDied","Data":"fbe39bc0c49faa915d2c4f50c757fc038107aae3b2ab6a7a97e1e90b02b1d1a4"} Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.725545 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-lk5mk" Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.725565 4775 scope.go:117] "RemoveContainer" containerID="180859970b041857dafe2a96bb332f8d1442cf9ef9839e6bbd7b85783c3dc75d" Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.731090 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cb53eb05-afa8-4039-805c-477b25ed3075","Type":"ContainerStarted","Data":"7645cb41a5451f5509943511872ab22f39410088b4a0322f6df0171a3ffb7fc5"} Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.731185 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cb53eb05-afa8-4039-805c-477b25ed3075","Type":"ContainerStarted","Data":"a2b7bcfad5b3fa6e886cea0c4b4a9a0a0cb851a1435a0c4faa718075cee23435"} Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.735919 4775 generic.go:334] "Generic (PLEG): container finished" podID="6b34a259-4754-41ea-b3b7-cd1ba5fb53c6" containerID="a5baf6934d04dd62ef1e7d5ae437ae7f3e8c190bcbdb11263300fe3c8b4015ff" exitCode=0 Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.736007 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-774db89647-9pnrf" event={"ID":"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6","Type":"ContainerDied","Data":"a5baf6934d04dd62ef1e7d5ae437ae7f3e8c190bcbdb11263300fe3c8b4015ff"} Mar 21 05:07:52 crc kubenswrapper[4775]: I0321 05:07:52.999188 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-lk5mk"] Mar 21 05:07:53 crc kubenswrapper[4775]: I0321 05:07:53.029187 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-lk5mk"] Mar 21 05:07:53 crc kubenswrapper[4775]: E0321 05:07:53.439984 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ac5c16c_56ac_4299_ae61_a8200986ce10.slice/crio-conmon-4e1e7c557d6fa58625eeced787347009a65ed5b8a3da8a6487f343ffff4e6c90.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ac5c16c_56ac_4299_ae61_a8200986ce10.slice/crio-4e1e7c557d6fa58625eeced787347009a65ed5b8a3da8a6487f343ffff4e6c90.scope\": RecentStats: unable to find data in memory cache]" Mar 21 05:07:53 crc kubenswrapper[4775]: I0321 05:07:53.648961 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 21 05:07:53 crc kubenswrapper[4775]: I0321 05:07:53.649057 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 21 05:07:53 crc kubenswrapper[4775]: I0321 05:07:53.676297 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f" path="/var/lib/kubelet/pods/f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f/volumes" Mar 21 05:07:53 crc kubenswrapper[4775]: I0321 05:07:53.681880 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 21 05:07:53 crc kubenswrapper[4775]: I0321 05:07:53.709048 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 21 05:07:53 crc kubenswrapper[4775]: I0321 05:07:53.748139 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cb53eb05-afa8-4039-805c-477b25ed3075","Type":"ContainerStarted","Data":"d8c073f8b7d47daaf28becd8c93adbf3ccc063178d35dc0cd6aeea148ed3060c"} Mar 21 05:07:53 crc kubenswrapper[4775]: I0321 05:07:53.748516 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 21 05:07:53 crc kubenswrapper[4775]: I0321 05:07:53.750286 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-774db89647-9pnrf" event={"ID":"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6","Type":"ContainerStarted","Data":"b889dfec32da8490ede4a5642cd397e9def74cf497031b982af0df0b06f79197"} Mar 21 05:07:53 crc kubenswrapper[4775]: I0321 05:07:53.750465 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:53 crc kubenswrapper[4775]: I0321 05:07:53.752789 4775 generic.go:334] "Generic (PLEG): container finished" podID="1658991f-7a2b-4ce9-a240-a940385e0b8f" containerID="796b087daa0094c51c318e24d71b74b2beb873149bcfc37bff7fc5e4bd3c8ed7" exitCode=0 Mar 21 05:07:53 crc kubenswrapper[4775]: I0321 05:07:53.752861 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9cmq2" event={"ID":"1658991f-7a2b-4ce9-a240-a940385e0b8f","Type":"ContainerDied","Data":"796b087daa0094c51c318e24d71b74b2beb873149bcfc37bff7fc5e4bd3c8ed7"} Mar 21 05:07:53 crc kubenswrapper[4775]: I0321 05:07:53.754569 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3f765864-79ee-415c-99be-55894a20c087","Type":"ContainerStarted","Data":"4764d465ee215948cf3a07d17ef04460952c808ab216c2d6600d30877448b571"} Mar 21 05:07:53 crc kubenswrapper[4775]: I0321 05:07:53.756711 4775 generic.go:334] "Generic (PLEG): container finished" podID="8ac5c16c-56ac-4299-ae61-a8200986ce10" containerID="4e1e7c557d6fa58625eeced787347009a65ed5b8a3da8a6487f343ffff4e6c90" exitCode=0 Mar 21 05:07:53 crc kubenswrapper[4775]: I0321 05:07:53.756802 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ac5c16c-56ac-4299-ae61-a8200986ce10","Type":"ContainerDied","Data":"4e1e7c557d6fa58625eeced787347009a65ed5b8a3da8a6487f343ffff4e6c90"} Mar 21 05:07:53 crc kubenswrapper[4775]: I0321 05:07:53.758600 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 21 05:07:53 crc kubenswrapper[4775]: I0321 05:07:53.758662 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 21 05:07:53 crc kubenswrapper[4775]: I0321 05:07:53.778771 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.778746196 podStartE2EDuration="3.778746196s" podCreationTimestamp="2026-03-21 05:07:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:07:53.769388261 +0000 UTC m=+1226.745851915" watchObservedRunningTime="2026-03-21 05:07:53.778746196 +0000 UTC m=+1226.755209820" Mar 21 05:07:53 crc kubenswrapper[4775]: I0321 05:07:53.831796 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-774db89647-9pnrf" podStartSLOduration=3.831772184 podStartE2EDuration="3.831772184s" podCreationTimestamp="2026-03-21 05:07:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:07:53.807865009 +0000 UTC m=+1226.784328633" watchObservedRunningTime="2026-03-21 05:07:53.831772184 +0000 UTC m=+1226.808235818" Mar 21 05:07:54 crc kubenswrapper[4775]: I0321 05:07:54.772953 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3f765864-79ee-415c-99be-55894a20c087","Type":"ContainerStarted","Data":"0b6150bc28d08d5b96097378ba0b885d4bf49a94a51c13823ebc2ecd1793e28b"} Mar 21 05:07:54 crc kubenswrapper[4775]: I0321 05:07:54.791518 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.967906501 podStartE2EDuration="4.791500357s" podCreationTimestamp="2026-03-21 05:07:50 +0000 UTC" firstStartedPulling="2026-03-21 05:07:51.41270204 +0000 UTC m=+1224.389165664" lastFinishedPulling="2026-03-21 05:07:52.236295896 +0000 UTC m=+1225.212759520" observedRunningTime="2026-03-21 05:07:54.791389804 +0000 UTC m=+1227.767853428" watchObservedRunningTime="2026-03-21 05:07:54.791500357 +0000 UTC m=+1227.767963981" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.211714 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9cmq2" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.262268 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bczb\" (UniqueName: \"kubernetes.io/projected/1658991f-7a2b-4ce9-a240-a940385e0b8f-kube-api-access-9bczb\") pod \"1658991f-7a2b-4ce9-a240-a940385e0b8f\" (UID: \"1658991f-7a2b-4ce9-a240-a940385e0b8f\") " Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.262554 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1658991f-7a2b-4ce9-a240-a940385e0b8f-db-sync-config-data\") pod \"1658991f-7a2b-4ce9-a240-a940385e0b8f\" (UID: \"1658991f-7a2b-4ce9-a240-a940385e0b8f\") " Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.262753 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1658991f-7a2b-4ce9-a240-a940385e0b8f-combined-ca-bundle\") pod \"1658991f-7a2b-4ce9-a240-a940385e0b8f\" (UID: \"1658991f-7a2b-4ce9-a240-a940385e0b8f\") " Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.270240 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1658991f-7a2b-4ce9-a240-a940385e0b8f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "1658991f-7a2b-4ce9-a240-a940385e0b8f" (UID: "1658991f-7a2b-4ce9-a240-a940385e0b8f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.271325 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1658991f-7a2b-4ce9-a240-a940385e0b8f-kube-api-access-9bczb" (OuterVolumeSpecName: "kube-api-access-9bczb") pod "1658991f-7a2b-4ce9-a240-a940385e0b8f" (UID: "1658991f-7a2b-4ce9-a240-a940385e0b8f"). InnerVolumeSpecName "kube-api-access-9bczb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.290373 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1658991f-7a2b-4ce9-a240-a940385e0b8f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1658991f-7a2b-4ce9-a240-a940385e0b8f" (UID: "1658991f-7a2b-4ce9-a240-a940385e0b8f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.364716 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bczb\" (UniqueName: \"kubernetes.io/projected/1658991f-7a2b-4ce9-a240-a940385e0b8f-kube-api-access-9bczb\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.364758 4775 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1658991f-7a2b-4ce9-a240-a940385e0b8f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.364769 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1658991f-7a2b-4ce9-a240-a940385e0b8f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.476373 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.590630 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7d65998c7c-prp5b"] Mar 21 05:07:55 crc kubenswrapper[4775]: E0321 05:07:55.591058 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1658991f-7a2b-4ce9-a240-a940385e0b8f" containerName="barbican-db-sync" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.591073 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1658991f-7a2b-4ce9-a240-a940385e0b8f" containerName="barbican-db-sync" Mar 21 05:07:55 crc kubenswrapper[4775]: E0321 05:07:55.591086 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f" containerName="init" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.591095 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f" containerName="init" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.591321 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1658991f-7a2b-4ce9-a240-a940385e0b8f" containerName="barbican-db-sync" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.591350 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5a4d8e0-d0d0-4b04-9f21-ddf0d94c0e8f" containerName="init" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.592393 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.595160 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.595948 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.596695 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7d65998c7c-prp5b"] Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.671495 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.674708 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-httpd-config\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.674771 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-internal-tls-certs\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.674813 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llxrd\" (UniqueName: \"kubernetes.io/projected/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-kube-api-access-llxrd\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.674874 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-public-tls-certs\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.674904 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-config\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.674950 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-combined-ca-bundle\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.674966 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-ovndb-tls-certs\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.734007 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.734061 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.770932 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.775897 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-combined-ca-bundle\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.775943 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-ovndb-tls-certs\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.775992 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-httpd-config\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.776044 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-internal-tls-certs\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.776076 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llxrd\" (UniqueName: \"kubernetes.io/projected/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-kube-api-access-llxrd\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.776170 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-public-tls-certs\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.776196 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-config\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.780019 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-internal-tls-certs\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.781874 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-ovndb-tls-certs\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.783535 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-combined-ca-bundle\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.784292 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-config\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.785070 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-public-tls-certs\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.787761 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-httpd-config\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.799698 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.805134 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llxrd\" (UniqueName: \"kubernetes.io/projected/3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef-kube-api-access-llxrd\") pod \"neutron-7d65998c7c-prp5b\" (UID: \"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef\") " pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.806550 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="cb53eb05-afa8-4039-805c-477b25ed3075" containerName="cinder-api-log" containerID="cri-o://7645cb41a5451f5509943511872ab22f39410088b4a0322f6df0171a3ffb7fc5" gracePeriod=30 Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.806823 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9cmq2" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.807461 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9cmq2" event={"ID":"1658991f-7a2b-4ce9-a240-a940385e0b8f","Type":"ContainerDied","Data":"cb42616105e59d6830a168b6da9f0c076bcf19cd878442d47ceb6cd33726aca0"} Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.807482 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb42616105e59d6830a168b6da9f0c076bcf19cd878442d47ceb6cd33726aca0" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.808719 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="cb53eb05-afa8-4039-805c-477b25ed3075" containerName="cinder-api" containerID="cri-o://d8c073f8b7d47daaf28becd8c93adbf3ccc063178d35dc0cd6aeea148ed3060c" gracePeriod=30 Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.808912 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.808927 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.913217 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.982481 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-75f5b547c8-mgjw5"] Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.984333 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.988480 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.989063 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Mar 21 05:07:55 crc kubenswrapper[4775]: I0321 05:07:55.989455 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-ph6zz" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.003387 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6dd9f89d55-fdf8c"] Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.005279 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6dd9f89d55-fdf8c" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.009020 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.045310 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6dd9f89d55-fdf8c"] Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.083342 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-75f5b547c8-mgjw5"] Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.161179 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-774db89647-9pnrf"] Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.161813 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-774db89647-9pnrf" podUID="6b34a259-4754-41ea-b3b7-cd1ba5fb53c6" containerName="dnsmasq-dns" containerID="cri-o://b889dfec32da8490ede4a5642cd397e9def74cf497031b982af0df0b06f79197" gracePeriod=10 Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.188062 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxt6r\" (UniqueName: \"kubernetes.io/projected/177228f6-7f69-49c2-9942-ea0a98b56d13-kube-api-access-hxt6r\") pod \"barbican-keystone-listener-75f5b547c8-mgjw5\" (UID: \"177228f6-7f69-49c2-9942-ea0a98b56d13\") " pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.188110 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/177228f6-7f69-49c2-9942-ea0a98b56d13-config-data-custom\") pod \"barbican-keystone-listener-75f5b547c8-mgjw5\" (UID: \"177228f6-7f69-49c2-9942-ea0a98b56d13\") " pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.188184 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8a7c2e5-3643-4675-9888-3c310e4f9ad4-config-data-custom\") pod \"barbican-worker-6dd9f89d55-fdf8c\" (UID: \"d8a7c2e5-3643-4675-9888-3c310e4f9ad4\") " pod="openstack/barbican-worker-6dd9f89d55-fdf8c" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.188224 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwjfh\" (UniqueName: \"kubernetes.io/projected/d8a7c2e5-3643-4675-9888-3c310e4f9ad4-kube-api-access-pwjfh\") pod \"barbican-worker-6dd9f89d55-fdf8c\" (UID: \"d8a7c2e5-3643-4675-9888-3c310e4f9ad4\") " pod="openstack/barbican-worker-6dd9f89d55-fdf8c" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.188255 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/177228f6-7f69-49c2-9942-ea0a98b56d13-logs\") pod \"barbican-keystone-listener-75f5b547c8-mgjw5\" (UID: \"177228f6-7f69-49c2-9942-ea0a98b56d13\") " pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.188279 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8a7c2e5-3643-4675-9888-3c310e4f9ad4-config-data\") pod \"barbican-worker-6dd9f89d55-fdf8c\" (UID: \"d8a7c2e5-3643-4675-9888-3c310e4f9ad4\") " pod="openstack/barbican-worker-6dd9f89d55-fdf8c" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.188319 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8a7c2e5-3643-4675-9888-3c310e4f9ad4-logs\") pod \"barbican-worker-6dd9f89d55-fdf8c\" (UID: \"d8a7c2e5-3643-4675-9888-3c310e4f9ad4\") " pod="openstack/barbican-worker-6dd9f89d55-fdf8c" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.188363 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/177228f6-7f69-49c2-9942-ea0a98b56d13-config-data\") pod \"barbican-keystone-listener-75f5b547c8-mgjw5\" (UID: \"177228f6-7f69-49c2-9942-ea0a98b56d13\") " pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.188387 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/177228f6-7f69-49c2-9942-ea0a98b56d13-combined-ca-bundle\") pod \"barbican-keystone-listener-75f5b547c8-mgjw5\" (UID: \"177228f6-7f69-49c2-9942-ea0a98b56d13\") " pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.188425 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8a7c2e5-3643-4675-9888-3c310e4f9ad4-combined-ca-bundle\") pod \"barbican-worker-6dd9f89d55-fdf8c\" (UID: \"d8a7c2e5-3643-4675-9888-3c310e4f9ad4\") " pod="openstack/barbican-worker-6dd9f89d55-fdf8c" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.243529 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-w2ltw"] Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.245007 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.289913 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxt6r\" (UniqueName: \"kubernetes.io/projected/177228f6-7f69-49c2-9942-ea0a98b56d13-kube-api-access-hxt6r\") pod \"barbican-keystone-listener-75f5b547c8-mgjw5\" (UID: \"177228f6-7f69-49c2-9942-ea0a98b56d13\") " pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.289962 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/177228f6-7f69-49c2-9942-ea0a98b56d13-config-data-custom\") pod \"barbican-keystone-listener-75f5b547c8-mgjw5\" (UID: \"177228f6-7f69-49c2-9942-ea0a98b56d13\") " pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.290006 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8a7c2e5-3643-4675-9888-3c310e4f9ad4-config-data-custom\") pod \"barbican-worker-6dd9f89d55-fdf8c\" (UID: \"d8a7c2e5-3643-4675-9888-3c310e4f9ad4\") " pod="openstack/barbican-worker-6dd9f89d55-fdf8c" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.290035 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwjfh\" (UniqueName: \"kubernetes.io/projected/d8a7c2e5-3643-4675-9888-3c310e4f9ad4-kube-api-access-pwjfh\") pod \"barbican-worker-6dd9f89d55-fdf8c\" (UID: \"d8a7c2e5-3643-4675-9888-3c310e4f9ad4\") " pod="openstack/barbican-worker-6dd9f89d55-fdf8c" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.290056 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/177228f6-7f69-49c2-9942-ea0a98b56d13-logs\") pod \"barbican-keystone-listener-75f5b547c8-mgjw5\" (UID: \"177228f6-7f69-49c2-9942-ea0a98b56d13\") " pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.290091 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8a7c2e5-3643-4675-9888-3c310e4f9ad4-config-data\") pod \"barbican-worker-6dd9f89d55-fdf8c\" (UID: \"d8a7c2e5-3643-4675-9888-3c310e4f9ad4\") " pod="openstack/barbican-worker-6dd9f89d55-fdf8c" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.290136 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8a7c2e5-3643-4675-9888-3c310e4f9ad4-logs\") pod \"barbican-worker-6dd9f89d55-fdf8c\" (UID: \"d8a7c2e5-3643-4675-9888-3c310e4f9ad4\") " pod="openstack/barbican-worker-6dd9f89d55-fdf8c" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.290179 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/177228f6-7f69-49c2-9942-ea0a98b56d13-config-data\") pod \"barbican-keystone-listener-75f5b547c8-mgjw5\" (UID: \"177228f6-7f69-49c2-9942-ea0a98b56d13\") " pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.290201 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/177228f6-7f69-49c2-9942-ea0a98b56d13-combined-ca-bundle\") pod \"barbican-keystone-listener-75f5b547c8-mgjw5\" (UID: \"177228f6-7f69-49c2-9942-ea0a98b56d13\") " pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.290230 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8a7c2e5-3643-4675-9888-3c310e4f9ad4-combined-ca-bundle\") pod \"barbican-worker-6dd9f89d55-fdf8c\" (UID: \"d8a7c2e5-3643-4675-9888-3c310e4f9ad4\") " pod="openstack/barbican-worker-6dd9f89d55-fdf8c" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.291794 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8a7c2e5-3643-4675-9888-3c310e4f9ad4-logs\") pod \"barbican-worker-6dd9f89d55-fdf8c\" (UID: \"d8a7c2e5-3643-4675-9888-3c310e4f9ad4\") " pod="openstack/barbican-worker-6dd9f89d55-fdf8c" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.292014 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/177228f6-7f69-49c2-9942-ea0a98b56d13-logs\") pod \"barbican-keystone-listener-75f5b547c8-mgjw5\" (UID: \"177228f6-7f69-49c2-9942-ea0a98b56d13\") " pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.295797 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-w2ltw"] Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.297007 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/177228f6-7f69-49c2-9942-ea0a98b56d13-config-data\") pod \"barbican-keystone-listener-75f5b547c8-mgjw5\" (UID: \"177228f6-7f69-49c2-9942-ea0a98b56d13\") " pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.305359 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/177228f6-7f69-49c2-9942-ea0a98b56d13-combined-ca-bundle\") pod \"barbican-keystone-listener-75f5b547c8-mgjw5\" (UID: \"177228f6-7f69-49c2-9942-ea0a98b56d13\") " pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.305444 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7dc8b6bf8b-btcqr"] Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.320869 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8a7c2e5-3643-4675-9888-3c310e4f9ad4-config-data\") pod \"barbican-worker-6dd9f89d55-fdf8c\" (UID: \"d8a7c2e5-3643-4675-9888-3c310e4f9ad4\") " pod="openstack/barbican-worker-6dd9f89d55-fdf8c" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.334735 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.337411 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.338638 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8a7c2e5-3643-4675-9888-3c310e4f9ad4-combined-ca-bundle\") pod \"barbican-worker-6dd9f89d55-fdf8c\" (UID: \"d8a7c2e5-3643-4675-9888-3c310e4f9ad4\") " pod="openstack/barbican-worker-6dd9f89d55-fdf8c" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.341260 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8a7c2e5-3643-4675-9888-3c310e4f9ad4-config-data-custom\") pod \"barbican-worker-6dd9f89d55-fdf8c\" (UID: \"d8a7c2e5-3643-4675-9888-3c310e4f9ad4\") " pod="openstack/barbican-worker-6dd9f89d55-fdf8c" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.344330 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwjfh\" (UniqueName: \"kubernetes.io/projected/d8a7c2e5-3643-4675-9888-3c310e4f9ad4-kube-api-access-pwjfh\") pod \"barbican-worker-6dd9f89d55-fdf8c\" (UID: \"d8a7c2e5-3643-4675-9888-3c310e4f9ad4\") " pod="openstack/barbican-worker-6dd9f89d55-fdf8c" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.345509 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7dc8b6bf8b-btcqr"] Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.350995 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/177228f6-7f69-49c2-9942-ea0a98b56d13-config-data-custom\") pod \"barbican-keystone-listener-75f5b547c8-mgjw5\" (UID: \"177228f6-7f69-49c2-9942-ea0a98b56d13\") " pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.366804 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6dd9f89d55-fdf8c" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.368995 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxt6r\" (UniqueName: \"kubernetes.io/projected/177228f6-7f69-49c2-9942-ea0a98b56d13-kube-api-access-hxt6r\") pod \"barbican-keystone-listener-75f5b547c8-mgjw5\" (UID: \"177228f6-7f69-49c2-9942-ea0a98b56d13\") " pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.392867 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-config\") pod \"dnsmasq-dns-6578955fd5-w2ltw\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.393141 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbc62\" (UniqueName: \"kubernetes.io/projected/d7dc929b-501f-43cb-8811-394841cc54f9-kube-api-access-tbc62\") pod \"dnsmasq-dns-6578955fd5-w2ltw\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.393181 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-dns-svc\") pod \"dnsmasq-dns-6578955fd5-w2ltw\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.393220 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-w2ltw\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.393280 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-w2ltw\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.393298 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-w2ltw\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.496727 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-config-data-custom\") pod \"barbican-api-7dc8b6bf8b-btcqr\" (UID: \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\") " pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.496772 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-config-data\") pod \"barbican-api-7dc8b6bf8b-btcqr\" (UID: \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\") " pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.496803 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-combined-ca-bundle\") pod \"barbican-api-7dc8b6bf8b-btcqr\" (UID: \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\") " pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.496831 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-w2ltw\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.496853 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-w2ltw\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.496914 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-config\") pod \"dnsmasq-dns-6578955fd5-w2ltw\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.496936 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbc62\" (UniqueName: \"kubernetes.io/projected/d7dc929b-501f-43cb-8811-394841cc54f9-kube-api-access-tbc62\") pod \"dnsmasq-dns-6578955fd5-w2ltw\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.496968 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-dns-svc\") pod \"dnsmasq-dns-6578955fd5-w2ltw\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.496983 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-logs\") pod \"barbican-api-7dc8b6bf8b-btcqr\" (UID: \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\") " pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.497029 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-w2ltw\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.497062 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bzmg\" (UniqueName: \"kubernetes.io/projected/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-kube-api-access-6bzmg\") pod \"barbican-api-7dc8b6bf8b-btcqr\" (UID: \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\") " pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.501041 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-w2ltw\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.501062 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-w2ltw\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.501816 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-dns-svc\") pod \"dnsmasq-dns-6578955fd5-w2ltw\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.501889 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-config\") pod \"dnsmasq-dns-6578955fd5-w2ltw\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.501832 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-w2ltw\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.591003 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbc62\" (UniqueName: \"kubernetes.io/projected/d7dc929b-501f-43cb-8811-394841cc54f9-kube-api-access-tbc62\") pod \"dnsmasq-dns-6578955fd5-w2ltw\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.598425 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-logs\") pod \"barbican-api-7dc8b6bf8b-btcqr\" (UID: \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\") " pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.599259 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bzmg\" (UniqueName: \"kubernetes.io/projected/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-kube-api-access-6bzmg\") pod \"barbican-api-7dc8b6bf8b-btcqr\" (UID: \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\") " pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.599412 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-config-data-custom\") pod \"barbican-api-7dc8b6bf8b-btcqr\" (UID: \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\") " pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.599519 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-config-data\") pod \"barbican-api-7dc8b6bf8b-btcqr\" (UID: \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\") " pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.599634 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-combined-ca-bundle\") pod \"barbican-api-7dc8b6bf8b-btcqr\" (UID: \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\") " pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.600298 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-logs\") pod \"barbican-api-7dc8b6bf8b-btcqr\" (UID: \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\") " pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.609951 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-config-data-custom\") pod \"barbican-api-7dc8b6bf8b-btcqr\" (UID: \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\") " pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.610989 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-combined-ca-bundle\") pod \"barbican-api-7dc8b6bf8b-btcqr\" (UID: \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\") " pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.612171 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-config-data\") pod \"barbican-api-7dc8b6bf8b-btcqr\" (UID: \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\") " pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.649796 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bzmg\" (UniqueName: \"kubernetes.io/projected/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-kube-api-access-6bzmg\") pod \"barbican-api-7dc8b6bf8b-btcqr\" (UID: \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\") " pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.660733 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.834971 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.847795 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.875515 4775 generic.go:334] "Generic (PLEG): container finished" podID="6b34a259-4754-41ea-b3b7-cd1ba5fb53c6" containerID="b889dfec32da8490ede4a5642cd397e9def74cf497031b982af0df0b06f79197" exitCode=0 Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.875651 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-774db89647-9pnrf" event={"ID":"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6","Type":"ContainerDied","Data":"b889dfec32da8490ede4a5642cd397e9def74cf497031b982af0df0b06f79197"} Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.897925 4775 generic.go:334] "Generic (PLEG): container finished" podID="cb53eb05-afa8-4039-805c-477b25ed3075" containerID="d8c073f8b7d47daaf28becd8c93adbf3ccc063178d35dc0cd6aeea148ed3060c" exitCode=0 Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.897955 4775 generic.go:334] "Generic (PLEG): container finished" podID="cb53eb05-afa8-4039-805c-477b25ed3075" containerID="7645cb41a5451f5509943511872ab22f39410088b4a0322f6df0171a3ffb7fc5" exitCode=143 Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.898849 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cb53eb05-afa8-4039-805c-477b25ed3075","Type":"ContainerDied","Data":"d8c073f8b7d47daaf28becd8c93adbf3ccc063178d35dc0cd6aeea148ed3060c"} Mar 21 05:07:56 crc kubenswrapper[4775]: I0321 05:07:56.898934 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cb53eb05-afa8-4039-805c-477b25ed3075","Type":"ContainerDied","Data":"7645cb41a5451f5509943511872ab22f39410088b4a0322f6df0171a3ffb7fc5"} Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.046473 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7d65998c7c-prp5b"] Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.054240 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.099975 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.114658 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-config-data-custom\") pod \"cb53eb05-afa8-4039-805c-477b25ed3075\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.114732 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctknn\" (UniqueName: \"kubernetes.io/projected/cb53eb05-afa8-4039-805c-477b25ed3075-kube-api-access-ctknn\") pod \"cb53eb05-afa8-4039-805c-477b25ed3075\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.114828 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-combined-ca-bundle\") pod \"cb53eb05-afa8-4039-805c-477b25ed3075\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.114887 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-scripts\") pod \"cb53eb05-afa8-4039-805c-477b25ed3075\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.114903 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb53eb05-afa8-4039-805c-477b25ed3075-logs\") pod \"cb53eb05-afa8-4039-805c-477b25ed3075\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.114934 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb53eb05-afa8-4039-805c-477b25ed3075-etc-machine-id\") pod \"cb53eb05-afa8-4039-805c-477b25ed3075\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.114961 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-config-data\") pod \"cb53eb05-afa8-4039-805c-477b25ed3075\" (UID: \"cb53eb05-afa8-4039-805c-477b25ed3075\") " Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.131657 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb53eb05-afa8-4039-805c-477b25ed3075-logs" (OuterVolumeSpecName: "logs") pod "cb53eb05-afa8-4039-805c-477b25ed3075" (UID: "cb53eb05-afa8-4039-805c-477b25ed3075"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.132101 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb53eb05-afa8-4039-805c-477b25ed3075-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "cb53eb05-afa8-4039-805c-477b25ed3075" (UID: "cb53eb05-afa8-4039-805c-477b25ed3075"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.135962 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cb53eb05-afa8-4039-805c-477b25ed3075" (UID: "cb53eb05-afa8-4039-805c-477b25ed3075"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.151470 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb53eb05-afa8-4039-805c-477b25ed3075-kube-api-access-ctknn" (OuterVolumeSpecName: "kube-api-access-ctknn") pod "cb53eb05-afa8-4039-805c-477b25ed3075" (UID: "cb53eb05-afa8-4039-805c-477b25ed3075"). InnerVolumeSpecName "kube-api-access-ctknn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.160354 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-scripts" (OuterVolumeSpecName: "scripts") pod "cb53eb05-afa8-4039-805c-477b25ed3075" (UID: "cb53eb05-afa8-4039-805c-477b25ed3075"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.214296 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb53eb05-afa8-4039-805c-477b25ed3075" (UID: "cb53eb05-afa8-4039-805c-477b25ed3075"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.216675 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-dns-swift-storage-0\") pod \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.216746 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4hqb\" (UniqueName: \"kubernetes.io/projected/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-kube-api-access-c4hqb\") pod \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.216862 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-ovsdbserver-sb\") pod \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.216916 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-dns-svc\") pod \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.216936 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-config\") pod \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.217028 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-ovsdbserver-nb\") pod \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\" (UID: \"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6\") " Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.217411 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.217421 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb53eb05-afa8-4039-805c-477b25ed3075-logs\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.217430 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.217438 4775 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb53eb05-afa8-4039-805c-477b25ed3075-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.217445 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.217454 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctknn\" (UniqueName: \"kubernetes.io/projected/cb53eb05-afa8-4039-805c-477b25ed3075-kube-api-access-ctknn\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.237535 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6dd9f89d55-fdf8c"] Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.251427 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-kube-api-access-c4hqb" (OuterVolumeSpecName: "kube-api-access-c4hqb") pod "6b34a259-4754-41ea-b3b7-cd1ba5fb53c6" (UID: "6b34a259-4754-41ea-b3b7-cd1ba5fb53c6"). InnerVolumeSpecName "kube-api-access-c4hqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.315737 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6b34a259-4754-41ea-b3b7-cd1ba5fb53c6" (UID: "6b34a259-4754-41ea-b3b7-cd1ba5fb53c6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.322008 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.322038 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4hqb\" (UniqueName: \"kubernetes.io/projected/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-kube-api-access-c4hqb\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.385556 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-config" (OuterVolumeSpecName: "config") pod "6b34a259-4754-41ea-b3b7-cd1ba5fb53c6" (UID: "6b34a259-4754-41ea-b3b7-cd1ba5fb53c6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.388999 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6b34a259-4754-41ea-b3b7-cd1ba5fb53c6" (UID: "6b34a259-4754-41ea-b3b7-cd1ba5fb53c6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.419969 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-config-data" (OuterVolumeSpecName: "config-data") pod "cb53eb05-afa8-4039-805c-477b25ed3075" (UID: "cb53eb05-afa8-4039-805c-477b25ed3075"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.424109 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb53eb05-afa8-4039-805c-477b25ed3075-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.424305 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.424317 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.448337 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-75f5b547c8-mgjw5"] Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.453688 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6b34a259-4754-41ea-b3b7-cd1ba5fb53c6" (UID: "6b34a259-4754-41ea-b3b7-cd1ba5fb53c6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.471867 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6b34a259-4754-41ea-b3b7-cd1ba5fb53c6" (UID: "6b34a259-4754-41ea-b3b7-cd1ba5fb53c6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.526309 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.526340 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.603959 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-w2ltw"] Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.639360 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7dc8b6bf8b-btcqr"] Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.908489 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" event={"ID":"177228f6-7f69-49c2-9942-ea0a98b56d13","Type":"ContainerStarted","Data":"d36081bb2a1cc1559310140250d86d1092b1275bdc8dd051dd3d2be34341636a"} Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.913690 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" event={"ID":"d7dc929b-501f-43cb-8811-394841cc54f9","Type":"ContainerStarted","Data":"4431b638594d2b34fefc401fb2d4af3e48aa80179d99ea9d13c2365707189fa3"} Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.918673 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" event={"ID":"86abc8a1-1d5e-4dad-9e62-4f1c88608b95","Type":"ContainerStarted","Data":"56894b1d2f57fb4749d6394cfec85a58eb1ab9e8c5e0205ea22bb5116fe8c5c9"} Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.924517 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cb53eb05-afa8-4039-805c-477b25ed3075","Type":"ContainerDied","Data":"a2b7bcfad5b3fa6e886cea0c4b4a9a0a0cb851a1435a0c4faa718075cee23435"} Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.924581 4775 scope.go:117] "RemoveContainer" containerID="d8c073f8b7d47daaf28becd8c93adbf3ccc063178d35dc0cd6aeea148ed3060c" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.924742 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.929937 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-774db89647-9pnrf" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.929949 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-774db89647-9pnrf" event={"ID":"6b34a259-4754-41ea-b3b7-cd1ba5fb53c6","Type":"ContainerDied","Data":"a73be0fd90826150202a28a361d434d06f770ab26ca7c2e2a7f4cd5e030488a3"} Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.933563 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6dd9f89d55-fdf8c" event={"ID":"d8a7c2e5-3643-4675-9888-3c310e4f9ad4","Type":"ContainerStarted","Data":"c196ea57d4dccfd615b126ec0536341650101ad3ff1e1ea6ab1956d5c98c8a4e"} Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.941246 4775 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.942698 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d65998c7c-prp5b" event={"ID":"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef","Type":"ContainerStarted","Data":"c6a873d5c7237282e58f0d52558bff33030d6b81e7787de55c1c7bc6d9dcd7b1"} Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.942754 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d65998c7c-prp5b" event={"ID":"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef","Type":"ContainerStarted","Data":"a2f9731ed7da4398238dad98c11c9d541a2a8817fc90ad06c6e43ea8fb25f8b3"} Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.975697 4775 scope.go:117] "RemoveContainer" containerID="7645cb41a5451f5509943511872ab22f39410088b4a0322f6df0171a3ffb7fc5" Mar 21 05:07:57 crc kubenswrapper[4775]: I0321 05:07:57.991176 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.002225 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.036167 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-774db89647-9pnrf"] Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.049965 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-774db89647-9pnrf"] Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.070331 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 21 05:07:58 crc kubenswrapper[4775]: E0321 05:07:58.070776 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb53eb05-afa8-4039-805c-477b25ed3075" containerName="cinder-api-log" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.070796 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb53eb05-afa8-4039-805c-477b25ed3075" containerName="cinder-api-log" Mar 21 05:07:58 crc kubenswrapper[4775]: E0321 05:07:58.070820 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b34a259-4754-41ea-b3b7-cd1ba5fb53c6" containerName="dnsmasq-dns" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.070830 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b34a259-4754-41ea-b3b7-cd1ba5fb53c6" containerName="dnsmasq-dns" Mar 21 05:07:58 crc kubenswrapper[4775]: E0321 05:07:58.070854 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b34a259-4754-41ea-b3b7-cd1ba5fb53c6" containerName="init" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.070862 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b34a259-4754-41ea-b3b7-cd1ba5fb53c6" containerName="init" Mar 21 05:07:58 crc kubenswrapper[4775]: E0321 05:07:58.070875 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb53eb05-afa8-4039-805c-477b25ed3075" containerName="cinder-api" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.070881 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb53eb05-afa8-4039-805c-477b25ed3075" containerName="cinder-api" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.071061 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb53eb05-afa8-4039-805c-477b25ed3075" containerName="cinder-api-log" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.071099 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb53eb05-afa8-4039-805c-477b25ed3075" containerName="cinder-api" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.071110 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b34a259-4754-41ea-b3b7-cd1ba5fb53c6" containerName="dnsmasq-dns" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.072298 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.101582 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.102220 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.111261 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.136308 4775 scope.go:117] "RemoveContainer" containerID="b889dfec32da8490ede4a5642cd397e9def74cf497031b982af0df0b06f79197" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.144152 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c16d835a-1ec2-473d-b2d8-c8e7c978e140-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.144210 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c16d835a-1ec2-473d-b2d8-c8e7c978e140-logs\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.144241 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c16d835a-1ec2-473d-b2d8-c8e7c978e140-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.144257 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16d835a-1ec2-473d-b2d8-c8e7c978e140-config-data\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.144311 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c16d835a-1ec2-473d-b2d8-c8e7c978e140-scripts\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.144333 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c16d835a-1ec2-473d-b2d8-c8e7c978e140-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.144354 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c16d835a-1ec2-473d-b2d8-c8e7c978e140-public-tls-certs\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.144377 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c16d835a-1ec2-473d-b2d8-c8e7c978e140-config-data-custom\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.144397 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvr69\" (UniqueName: \"kubernetes.io/projected/c16d835a-1ec2-473d-b2d8-c8e7c978e140-kube-api-access-nvr69\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.152432 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.202447 4775 scope.go:117] "RemoveContainer" containerID="a5baf6934d04dd62ef1e7d5ae437ae7f3e8c190bcbdb11263300fe3c8b4015ff" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.245999 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c16d835a-1ec2-473d-b2d8-c8e7c978e140-logs\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.246060 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c16d835a-1ec2-473d-b2d8-c8e7c978e140-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.246082 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16d835a-1ec2-473d-b2d8-c8e7c978e140-config-data\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.250291 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c16d835a-1ec2-473d-b2d8-c8e7c978e140-scripts\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.250380 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c16d835a-1ec2-473d-b2d8-c8e7c978e140-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.250439 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c16d835a-1ec2-473d-b2d8-c8e7c978e140-public-tls-certs\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.250499 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c16d835a-1ec2-473d-b2d8-c8e7c978e140-config-data-custom\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.250539 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvr69\" (UniqueName: \"kubernetes.io/projected/c16d835a-1ec2-473d-b2d8-c8e7c978e140-kube-api-access-nvr69\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.250699 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c16d835a-1ec2-473d-b2d8-c8e7c978e140-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.250904 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c16d835a-1ec2-473d-b2d8-c8e7c978e140-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.253729 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c16d835a-1ec2-473d-b2d8-c8e7c978e140-logs\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.261002 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c16d835a-1ec2-473d-b2d8-c8e7c978e140-config-data-custom\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.263443 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c16d835a-1ec2-473d-b2d8-c8e7c978e140-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.271915 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c16d835a-1ec2-473d-b2d8-c8e7c978e140-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.272080 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c16d835a-1ec2-473d-b2d8-c8e7c978e140-public-tls-certs\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.273562 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvr69\" (UniqueName: \"kubernetes.io/projected/c16d835a-1ec2-473d-b2d8-c8e7c978e140-kube-api-access-nvr69\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.273790 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16d835a-1ec2-473d-b2d8-c8e7c978e140-config-data\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.285707 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c16d835a-1ec2-473d-b2d8-c8e7c978e140-scripts\") pod \"cinder-api-0\" (UID: \"c16d835a-1ec2-473d-b2d8-c8e7c978e140\") " pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.501814 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.555767 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:07:58 crc kubenswrapper[4775]: I0321 05:07:58.660017 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:07:59 crc kubenswrapper[4775]: I0321 05:07:58.997146 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d65998c7c-prp5b" event={"ID":"3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef","Type":"ContainerStarted","Data":"66d5512823bb0c671e4b2e3db5aaa6a1ed79a9451ed17faf46bfaddedc2a672c"} Mar 21 05:07:59 crc kubenswrapper[4775]: I0321 05:07:58.997447 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:07:59 crc kubenswrapper[4775]: I0321 05:07:59.021912 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7d65998c7c-prp5b" podStartSLOduration=4.02189394 podStartE2EDuration="4.02189394s" podCreationTimestamp="2026-03-21 05:07:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:07:59.018924056 +0000 UTC m=+1231.995387680" watchObservedRunningTime="2026-03-21 05:07:59.02189394 +0000 UTC m=+1231.998357554" Mar 21 05:07:59 crc kubenswrapper[4775]: I0321 05:07:59.022141 4775 generic.go:334] "Generic (PLEG): container finished" podID="d7dc929b-501f-43cb-8811-394841cc54f9" containerID="6ab789fb8fabdc5b2c75b49b35b369108626a6a98dbf396798f603e25b157d25" exitCode=0 Mar 21 05:07:59 crc kubenswrapper[4775]: I0321 05:07:59.022207 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" event={"ID":"d7dc929b-501f-43cb-8811-394841cc54f9","Type":"ContainerDied","Data":"6ab789fb8fabdc5b2c75b49b35b369108626a6a98dbf396798f603e25b157d25"} Mar 21 05:07:59 crc kubenswrapper[4775]: I0321 05:07:59.029171 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" event={"ID":"86abc8a1-1d5e-4dad-9e62-4f1c88608b95","Type":"ContainerStarted","Data":"9e09a5090c4d61aabf6c91b2097f27d3892ae005a5644d72b8acf7ad49174027"} Mar 21 05:07:59 crc kubenswrapper[4775]: I0321 05:07:59.029407 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" event={"ID":"86abc8a1-1d5e-4dad-9e62-4f1c88608b95","Type":"ContainerStarted","Data":"da40033a5dd52e534fbf0e897d41a379c55c497ad81da6c097f013d7acf158d3"} Mar 21 05:07:59 crc kubenswrapper[4775]: I0321 05:07:59.030010 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:07:59 crc kubenswrapper[4775]: I0321 05:07:59.030150 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:07:59 crc kubenswrapper[4775]: I0321 05:07:59.098609 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" podStartSLOduration=3.098588838 podStartE2EDuration="3.098588838s" podCreationTimestamp="2026-03-21 05:07:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:07:59.078074758 +0000 UTC m=+1232.054538382" watchObservedRunningTime="2026-03-21 05:07:59.098588838 +0000 UTC m=+1232.075052462" Mar 21 05:07:59 crc kubenswrapper[4775]: I0321 05:07:59.150688 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 21 05:07:59 crc kubenswrapper[4775]: I0321 05:07:59.234892 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 21 05:07:59 crc kubenswrapper[4775]: I0321 05:07:59.235086 4775 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 21 05:07:59 crc kubenswrapper[4775]: I0321 05:07:59.676665 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b34a259-4754-41ea-b3b7-cd1ba5fb53c6" path="/var/lib/kubelet/pods/6b34a259-4754-41ea-b3b7-cd1ba5fb53c6/volumes" Mar 21 05:07:59 crc kubenswrapper[4775]: I0321 05:07:59.677817 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb53eb05-afa8-4039-805c-477b25ed3075" path="/var/lib/kubelet/pods/cb53eb05-afa8-4039-805c-477b25ed3075/volumes" Mar 21 05:07:59 crc kubenswrapper[4775]: I0321 05:07:59.974582 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.072584 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-77fd86567d-mf2wb"] Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.075995 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.078597 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.078787 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.090339 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c16d835a-1ec2-473d-b2d8-c8e7c978e140","Type":"ContainerStarted","Data":"e7f287bd42031f92358435f1c28406a556f8b03cbd50eec4d7c178aa7bbaac92"} Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.095908 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-77fd86567d-mf2wb"] Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.113034 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" event={"ID":"d7dc929b-501f-43cb-8811-394841cc54f9","Type":"ContainerStarted","Data":"07d83e49b0e806d759a2ee326367b45aab2ed5d87b3ee996d3b75ec5f20c736e"} Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.113848 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.141727 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" podStartSLOduration=4.141704707 podStartE2EDuration="4.141704707s" podCreationTimestamp="2026-03-21 05:07:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:08:00.141567623 +0000 UTC m=+1233.118031247" watchObservedRunningTime="2026-03-21 05:08:00.141704707 +0000 UTC m=+1233.118168331" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.178200 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567828-6klnl"] Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.179512 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567828-6klnl" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.182492 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.182666 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.183181 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.192435 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1727040d-36f5-431c-b8f1-84e206146dcf-config-data\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.192492 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1727040d-36f5-431c-b8f1-84e206146dcf-internal-tls-certs\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.192522 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1727040d-36f5-431c-b8f1-84e206146dcf-logs\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.192551 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1727040d-36f5-431c-b8f1-84e206146dcf-config-data-custom\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.192579 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49fqs\" (UniqueName: \"kubernetes.io/projected/1727040d-36f5-431c-b8f1-84e206146dcf-kube-api-access-49fqs\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.192597 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1727040d-36f5-431c-b8f1-84e206146dcf-combined-ca-bundle\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.192675 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1727040d-36f5-431c-b8f1-84e206146dcf-public-tls-certs\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.194500 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567828-6klnl"] Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.294246 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1727040d-36f5-431c-b8f1-84e206146dcf-internal-tls-certs\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.294341 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1727040d-36f5-431c-b8f1-84e206146dcf-logs\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.294432 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1727040d-36f5-431c-b8f1-84e206146dcf-config-data-custom\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.294484 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49fqs\" (UniqueName: \"kubernetes.io/projected/1727040d-36f5-431c-b8f1-84e206146dcf-kube-api-access-49fqs\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.294523 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1727040d-36f5-431c-b8f1-84e206146dcf-combined-ca-bundle\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.294559 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gt4qg\" (UniqueName: \"kubernetes.io/projected/eccf102b-f91a-4251-b878-1c21eea92522-kube-api-access-gt4qg\") pod \"auto-csr-approver-29567828-6klnl\" (UID: \"eccf102b-f91a-4251-b878-1c21eea92522\") " pod="openshift-infra/auto-csr-approver-29567828-6klnl" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.294779 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1727040d-36f5-431c-b8f1-84e206146dcf-public-tls-certs\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.294891 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1727040d-36f5-431c-b8f1-84e206146dcf-config-data\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.296393 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1727040d-36f5-431c-b8f1-84e206146dcf-logs\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.300629 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1727040d-36f5-431c-b8f1-84e206146dcf-internal-tls-certs\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.304777 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1727040d-36f5-431c-b8f1-84e206146dcf-combined-ca-bundle\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.304869 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1727040d-36f5-431c-b8f1-84e206146dcf-config-data\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.325153 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1727040d-36f5-431c-b8f1-84e206146dcf-public-tls-certs\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.352714 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49fqs\" (UniqueName: \"kubernetes.io/projected/1727040d-36f5-431c-b8f1-84e206146dcf-kube-api-access-49fqs\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.386158 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1727040d-36f5-431c-b8f1-84e206146dcf-config-data-custom\") pod \"barbican-api-77fd86567d-mf2wb\" (UID: \"1727040d-36f5-431c-b8f1-84e206146dcf\") " pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.398216 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gt4qg\" (UniqueName: \"kubernetes.io/projected/eccf102b-f91a-4251-b878-1c21eea92522-kube-api-access-gt4qg\") pod \"auto-csr-approver-29567828-6klnl\" (UID: \"eccf102b-f91a-4251-b878-1c21eea92522\") " pod="openshift-infra/auto-csr-approver-29567828-6klnl" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.401548 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.439436 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gt4qg\" (UniqueName: \"kubernetes.io/projected/eccf102b-f91a-4251-b878-1c21eea92522-kube-api-access-gt4qg\") pod \"auto-csr-approver-29567828-6klnl\" (UID: \"eccf102b-f91a-4251-b878-1c21eea92522\") " pod="openshift-infra/auto-csr-approver-29567828-6klnl" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.512620 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567828-6klnl" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.874459 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6496ddbdd4-v5mc5" Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.959361 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-64fb567758-hd2ld"] Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.959781 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-64fb567758-hd2ld" podUID="09cf3763-0b41-4452-a247-d9a56f58b05d" containerName="horizon-log" containerID="cri-o://4df1223c20d1958e562bffafa09f6a9587b632ec38a4f1dbf6a26792e8ce21fd" gracePeriod=30 Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.959980 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-64fb567758-hd2ld" podUID="09cf3763-0b41-4452-a247-d9a56f58b05d" containerName="horizon" containerID="cri-o://1da6670625f81794dc89cffb06ac6d008a9c7c8ed1b16d49c09b2332ff0e94fa" gracePeriod=30 Mar 21 05:08:00 crc kubenswrapper[4775]: I0321 05:08:00.976596 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-64fb567758-hd2ld" podUID="09cf3763-0b41-4452-a247-d9a56f58b05d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Mar 21 05:08:01 crc kubenswrapper[4775]: I0321 05:08:01.012708 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 21 05:08:01 crc kubenswrapper[4775]: I0321 05:08:01.051973 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 21 05:08:01 crc kubenswrapper[4775]: I0321 05:08:01.121763 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c16d835a-1ec2-473d-b2d8-c8e7c978e140","Type":"ContainerStarted","Data":"7e429c498fc0f4914500c6d809a65a750f1778d61663094beaa9607086649ce5"} Mar 21 05:08:01 crc kubenswrapper[4775]: I0321 05:08:01.121841 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="3f765864-79ee-415c-99be-55894a20c087" containerName="cinder-scheduler" containerID="cri-o://4764d465ee215948cf3a07d17ef04460952c808ab216c2d6600d30877448b571" gracePeriod=30 Mar 21 05:08:01 crc kubenswrapper[4775]: I0321 05:08:01.121900 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="3f765864-79ee-415c-99be-55894a20c087" containerName="probe" containerID="cri-o://0b6150bc28d08d5b96097378ba0b885d4bf49a94a51c13823ebc2ecd1793e28b" gracePeriod=30 Mar 21 05:08:01 crc kubenswrapper[4775]: I0321 05:08:01.651359 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-77fd86567d-mf2wb"] Mar 21 05:08:01 crc kubenswrapper[4775]: W0321 05:08:01.675221 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1727040d_36f5_431c_b8f1_84e206146dcf.slice/crio-5f510ceba77ac4c1d3cb0ef8dfe2b2f74dd976f3a7826bebbd6a3e86f98c64b7 WatchSource:0}: Error finding container 5f510ceba77ac4c1d3cb0ef8dfe2b2f74dd976f3a7826bebbd6a3e86f98c64b7: Status 404 returned error can't find the container with id 5f510ceba77ac4c1d3cb0ef8dfe2b2f74dd976f3a7826bebbd6a3e86f98c64b7 Mar 21 05:08:01 crc kubenswrapper[4775]: I0321 05:08:01.722089 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567828-6klnl"] Mar 21 05:08:01 crc kubenswrapper[4775]: W0321 05:08:01.758095 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeccf102b_f91a_4251_b878_1c21eea92522.slice/crio-486b3020022220622c2fb587ef965adb79954d3ca5c1cadb0f004636ab8a7a5f WatchSource:0}: Error finding container 486b3020022220622c2fb587ef965adb79954d3ca5c1cadb0f004636ab8a7a5f: Status 404 returned error can't find the container with id 486b3020022220622c2fb587ef965adb79954d3ca5c1cadb0f004636ab8a7a5f Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.133655 4775 generic.go:334] "Generic (PLEG): container finished" podID="3f765864-79ee-415c-99be-55894a20c087" containerID="0b6150bc28d08d5b96097378ba0b885d4bf49a94a51c13823ebc2ecd1793e28b" exitCode=0 Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.133727 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3f765864-79ee-415c-99be-55894a20c087","Type":"ContainerDied","Data":"0b6150bc28d08d5b96097378ba0b885d4bf49a94a51c13823ebc2ecd1793e28b"} Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.135443 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567828-6klnl" event={"ID":"eccf102b-f91a-4251-b878-1c21eea92522","Type":"ContainerStarted","Data":"486b3020022220622c2fb587ef965adb79954d3ca5c1cadb0f004636ab8a7a5f"} Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.137462 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" event={"ID":"177228f6-7f69-49c2-9942-ea0a98b56d13","Type":"ContainerStarted","Data":"a4f1a7b8252d5d938c9f37a5c854f86c11cfca913cfd03013441dcd3d7e9a4b0"} Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.137509 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" event={"ID":"177228f6-7f69-49c2-9942-ea0a98b56d13","Type":"ContainerStarted","Data":"290c55a6ed9a3ccda29d74736c7303706a061c58f9cfbcd56bca73fbde4d0be1"} Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.140129 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c16d835a-1ec2-473d-b2d8-c8e7c978e140","Type":"ContainerStarted","Data":"95c44b78dbc399f3bb0547a6982db27b3eb10d48a82b05ae66a24f89138c5a4a"} Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.140260 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.142403 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-77fd86567d-mf2wb" event={"ID":"1727040d-36f5-431c-b8f1-84e206146dcf","Type":"ContainerStarted","Data":"89930651d1c2dbb77788af4c417745e4d0c36ec1ff1f256a634a6fc979de8740"} Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.142468 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-77fd86567d-mf2wb" event={"ID":"1727040d-36f5-431c-b8f1-84e206146dcf","Type":"ContainerStarted","Data":"77c5ea4c20c4d779f476066bc96c8a9d0dc23e794f7f3d5e06ca5c68bd151774"} Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.142487 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-77fd86567d-mf2wb" event={"ID":"1727040d-36f5-431c-b8f1-84e206146dcf","Type":"ContainerStarted","Data":"5f510ceba77ac4c1d3cb0ef8dfe2b2f74dd976f3a7826bebbd6a3e86f98c64b7"} Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.142593 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.144662 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6dd9f89d55-fdf8c" event={"ID":"d8a7c2e5-3643-4675-9888-3c310e4f9ad4","Type":"ContainerStarted","Data":"5ada761d9c2f07a9ac6841539f8838edebf3297660a3e8de73a390e936d49876"} Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.144871 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6dd9f89d55-fdf8c" event={"ID":"d8a7c2e5-3643-4675-9888-3c310e4f9ad4","Type":"ContainerStarted","Data":"0c9ad89e2e596eca72945de953e3d9cd64eec0394196ab36a4b90dd4fabadeab"} Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.163842 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-75f5b547c8-mgjw5" podStartSLOduration=3.422676166 podStartE2EDuration="7.163824363s" podCreationTimestamp="2026-03-21 05:07:55 +0000 UTC" firstStartedPulling="2026-03-21 05:07:57.452286362 +0000 UTC m=+1230.428749986" lastFinishedPulling="2026-03-21 05:08:01.193434559 +0000 UTC m=+1234.169898183" observedRunningTime="2026-03-21 05:08:02.156598549 +0000 UTC m=+1235.133062173" watchObservedRunningTime="2026-03-21 05:08:02.163824363 +0000 UTC m=+1235.140287987" Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.183817 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.183789047 podStartE2EDuration="4.183789047s" podCreationTimestamp="2026-03-21 05:07:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:08:02.174951387 +0000 UTC m=+1235.151415011" watchObservedRunningTime="2026-03-21 05:08:02.183789047 +0000 UTC m=+1235.160252691" Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.196365 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6dd9f89d55-fdf8c" podStartSLOduration=3.247045353 podStartE2EDuration="7.196341632s" podCreationTimestamp="2026-03-21 05:07:55 +0000 UTC" firstStartedPulling="2026-03-21 05:07:57.243894183 +0000 UTC m=+1230.220357807" lastFinishedPulling="2026-03-21 05:08:01.193190452 +0000 UTC m=+1234.169654086" observedRunningTime="2026-03-21 05:08:02.193973235 +0000 UTC m=+1235.170436859" watchObservedRunningTime="2026-03-21 05:08:02.196341632 +0000 UTC m=+1235.172805266" Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.481980 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.482307 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.482348 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.483026 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"09c71a6e96dc622a58adf5b83a67eab26ff45301d2bfcc2a43f1cbd9eb2d9791"} pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 05:08:02 crc kubenswrapper[4775]: I0321 05:08:02.483097 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" containerID="cri-o://09c71a6e96dc622a58adf5b83a67eab26ff45301d2bfcc2a43f1cbd9eb2d9791" gracePeriod=600 Mar 21 05:08:03 crc kubenswrapper[4775]: I0321 05:08:03.173975 4775 generic.go:334] "Generic (PLEG): container finished" podID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerID="09c71a6e96dc622a58adf5b83a67eab26ff45301d2bfcc2a43f1cbd9eb2d9791" exitCode=0 Mar 21 05:08:03 crc kubenswrapper[4775]: I0321 05:08:03.174085 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerDied","Data":"09c71a6e96dc622a58adf5b83a67eab26ff45301d2bfcc2a43f1cbd9eb2d9791"} Mar 21 05:08:03 crc kubenswrapper[4775]: I0321 05:08:03.176375 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerStarted","Data":"b6e85a7c4acc97394b06df812a396284f471ec9c7f9eee22918e9da54e21feda"} Mar 21 05:08:03 crc kubenswrapper[4775]: I0321 05:08:03.176399 4775 scope.go:117] "RemoveContainer" containerID="8d303e1558b2adccb16582693a30248fb7f96ca561d7dcb3104197e825dd15a7" Mar 21 05:08:03 crc kubenswrapper[4775]: I0321 05:08:03.179158 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:03 crc kubenswrapper[4775]: I0321 05:08:03.199592 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-77fd86567d-mf2wb" podStartSLOduration=3.199525533 podStartE2EDuration="3.199525533s" podCreationTimestamp="2026-03-21 05:08:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:08:02.223273253 +0000 UTC m=+1235.199736877" watchObservedRunningTime="2026-03-21 05:08:03.199525533 +0000 UTC m=+1236.175989157" Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.190786 4775 generic.go:334] "Generic (PLEG): container finished" podID="3f765864-79ee-415c-99be-55894a20c087" containerID="4764d465ee215948cf3a07d17ef04460952c808ab216c2d6600d30877448b571" exitCode=0 Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.193620 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3f765864-79ee-415c-99be-55894a20c087","Type":"ContainerDied","Data":"4764d465ee215948cf3a07d17ef04460952c808ab216c2d6600d30877448b571"} Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.196431 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567828-6klnl" event={"ID":"eccf102b-f91a-4251-b878-1c21eea92522","Type":"ContainerStarted","Data":"bca1df983777352aa51a410cf20699daccf33fb2338a4057481b7410153afd88"} Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.219199 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29567828-6klnl" podStartSLOduration=2.298112147 podStartE2EDuration="4.219180948s" podCreationTimestamp="2026-03-21 05:08:00 +0000 UTC" firstStartedPulling="2026-03-21 05:08:01.760629588 +0000 UTC m=+1234.737093212" lastFinishedPulling="2026-03-21 05:08:03.681698389 +0000 UTC m=+1236.658162013" observedRunningTime="2026-03-21 05:08:04.217036397 +0000 UTC m=+1237.193500021" watchObservedRunningTime="2026-03-21 05:08:04.219180948 +0000 UTC m=+1237.195644582" Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.599663 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.681419 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-scripts\") pod \"3f765864-79ee-415c-99be-55894a20c087\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.681477 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-config-data\") pod \"3f765864-79ee-415c-99be-55894a20c087\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.681512 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-combined-ca-bundle\") pod \"3f765864-79ee-415c-99be-55894a20c087\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.681554 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-config-data-custom\") pod \"3f765864-79ee-415c-99be-55894a20c087\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.681964 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f765864-79ee-415c-99be-55894a20c087-etc-machine-id\") pod \"3f765864-79ee-415c-99be-55894a20c087\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.682009 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krwll\" (UniqueName: \"kubernetes.io/projected/3f765864-79ee-415c-99be-55894a20c087-kube-api-access-krwll\") pod \"3f765864-79ee-415c-99be-55894a20c087\" (UID: \"3f765864-79ee-415c-99be-55894a20c087\") " Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.687765 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f765864-79ee-415c-99be-55894a20c087-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3f765864-79ee-415c-99be-55894a20c087" (UID: "3f765864-79ee-415c-99be-55894a20c087"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.690363 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3f765864-79ee-415c-99be-55894a20c087" (UID: "3f765864-79ee-415c-99be-55894a20c087"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.690666 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f765864-79ee-415c-99be-55894a20c087-kube-api-access-krwll" (OuterVolumeSpecName: "kube-api-access-krwll") pod "3f765864-79ee-415c-99be-55894a20c087" (UID: "3f765864-79ee-415c-99be-55894a20c087"). InnerVolumeSpecName "kube-api-access-krwll". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.692936 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-scripts" (OuterVolumeSpecName: "scripts") pod "3f765864-79ee-415c-99be-55894a20c087" (UID: "3f765864-79ee-415c-99be-55894a20c087"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.762955 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f765864-79ee-415c-99be-55894a20c087" (UID: "3f765864-79ee-415c-99be-55894a20c087"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.783976 4775 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f765864-79ee-415c-99be-55894a20c087-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.784013 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krwll\" (UniqueName: \"kubernetes.io/projected/3f765864-79ee-415c-99be-55894a20c087-kube-api-access-krwll\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.784028 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.784040 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.784052 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.796485 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-config-data" (OuterVolumeSpecName: "config-data") pod "3f765864-79ee-415c-99be-55894a20c087" (UID: "3f765864-79ee-415c-99be-55894a20c087"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:04 crc kubenswrapper[4775]: I0321 05:08:04.886013 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f765864-79ee-415c-99be-55894a20c087-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.224824 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3f765864-79ee-415c-99be-55894a20c087","Type":"ContainerDied","Data":"85e53e5965bf199e0c4123a3954f4e5d437587f60519d3c96ac82b5f50bb3679"} Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.224863 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.225364 4775 scope.go:117] "RemoveContainer" containerID="0b6150bc28d08d5b96097378ba0b885d4bf49a94a51c13823ebc2ecd1793e28b" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.230212 4775 generic.go:334] "Generic (PLEG): container finished" podID="eccf102b-f91a-4251-b878-1c21eea92522" containerID="bca1df983777352aa51a410cf20699daccf33fb2338a4057481b7410153afd88" exitCode=0 Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.230276 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567828-6klnl" event={"ID":"eccf102b-f91a-4251-b878-1c21eea92522","Type":"ContainerDied","Data":"bca1df983777352aa51a410cf20699daccf33fb2338a4057481b7410153afd88"} Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.262828 4775 scope.go:117] "RemoveContainer" containerID="4764d465ee215948cf3a07d17ef04460952c808ab216c2d6600d30877448b571" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.275763 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.295855 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.308973 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 21 05:08:05 crc kubenswrapper[4775]: E0321 05:08:05.309426 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f765864-79ee-415c-99be-55894a20c087" containerName="probe" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.309439 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f765864-79ee-415c-99be-55894a20c087" containerName="probe" Mar 21 05:08:05 crc kubenswrapper[4775]: E0321 05:08:05.309465 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f765864-79ee-415c-99be-55894a20c087" containerName="cinder-scheduler" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.309472 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f765864-79ee-415c-99be-55894a20c087" containerName="cinder-scheduler" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.309633 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f765864-79ee-415c-99be-55894a20c087" containerName="cinder-scheduler" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.309660 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f765864-79ee-415c-99be-55894a20c087" containerName="probe" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.310617 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.312743 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.338015 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.395975 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d02bb319-292e-449c-8e5b-42c6859f1529-config-data\") pod \"cinder-scheduler-0\" (UID: \"d02bb319-292e-449c-8e5b-42c6859f1529\") " pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.396017 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsqx6\" (UniqueName: \"kubernetes.io/projected/d02bb319-292e-449c-8e5b-42c6859f1529-kube-api-access-vsqx6\") pod \"cinder-scheduler-0\" (UID: \"d02bb319-292e-449c-8e5b-42c6859f1529\") " pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.396357 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d02bb319-292e-449c-8e5b-42c6859f1529-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d02bb319-292e-449c-8e5b-42c6859f1529\") " pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.396442 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d02bb319-292e-449c-8e5b-42c6859f1529-scripts\") pod \"cinder-scheduler-0\" (UID: \"d02bb319-292e-449c-8e5b-42c6859f1529\") " pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.396476 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d02bb319-292e-449c-8e5b-42c6859f1529-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d02bb319-292e-449c-8e5b-42c6859f1529\") " pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.396541 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02bb319-292e-449c-8e5b-42c6859f1529-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d02bb319-292e-449c-8e5b-42c6859f1529\") " pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.399307 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-64fb567758-hd2ld" podUID="09cf3763-0b41-4452-a247-d9a56f58b05d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:57804->10.217.0.147:8443: read: connection reset by peer" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.399907 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-64fb567758-hd2ld" podUID="09cf3763-0b41-4452-a247-d9a56f58b05d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.499307 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02bb319-292e-449c-8e5b-42c6859f1529-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d02bb319-292e-449c-8e5b-42c6859f1529\") " pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.499433 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d02bb319-292e-449c-8e5b-42c6859f1529-config-data\") pod \"cinder-scheduler-0\" (UID: \"d02bb319-292e-449c-8e5b-42c6859f1529\") " pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.499463 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsqx6\" (UniqueName: \"kubernetes.io/projected/d02bb319-292e-449c-8e5b-42c6859f1529-kube-api-access-vsqx6\") pod \"cinder-scheduler-0\" (UID: \"d02bb319-292e-449c-8e5b-42c6859f1529\") " pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.499591 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d02bb319-292e-449c-8e5b-42c6859f1529-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d02bb319-292e-449c-8e5b-42c6859f1529\") " pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.499618 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d02bb319-292e-449c-8e5b-42c6859f1529-scripts\") pod \"cinder-scheduler-0\" (UID: \"d02bb319-292e-449c-8e5b-42c6859f1529\") " pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.499640 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d02bb319-292e-449c-8e5b-42c6859f1529-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d02bb319-292e-449c-8e5b-42c6859f1529\") " pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.499750 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d02bb319-292e-449c-8e5b-42c6859f1529-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d02bb319-292e-449c-8e5b-42c6859f1529\") " pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.505169 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02bb319-292e-449c-8e5b-42c6859f1529-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d02bb319-292e-449c-8e5b-42c6859f1529\") " pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.508169 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d02bb319-292e-449c-8e5b-42c6859f1529-config-data\") pod \"cinder-scheduler-0\" (UID: \"d02bb319-292e-449c-8e5b-42c6859f1529\") " pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.508532 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d02bb319-292e-449c-8e5b-42c6859f1529-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d02bb319-292e-449c-8e5b-42c6859f1529\") " pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.510517 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d02bb319-292e-449c-8e5b-42c6859f1529-scripts\") pod \"cinder-scheduler-0\" (UID: \"d02bb319-292e-449c-8e5b-42c6859f1529\") " pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.520771 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsqx6\" (UniqueName: \"kubernetes.io/projected/d02bb319-292e-449c-8e5b-42c6859f1529-kube-api-access-vsqx6\") pod \"cinder-scheduler-0\" (UID: \"d02bb319-292e-449c-8e5b-42c6859f1529\") " pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.680692 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 21 05:08:05 crc kubenswrapper[4775]: I0321 05:08:05.682891 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f765864-79ee-415c-99be-55894a20c087" path="/var/lib/kubelet/pods/3f765864-79ee-415c-99be-55894a20c087/volumes" Mar 21 05:08:06 crc kubenswrapper[4775]: I0321 05:08:06.175144 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 21 05:08:06 crc kubenswrapper[4775]: I0321 05:08:06.249735 4775 generic.go:334] "Generic (PLEG): container finished" podID="09cf3763-0b41-4452-a247-d9a56f58b05d" containerID="1da6670625f81794dc89cffb06ac6d008a9c7c8ed1b16d49c09b2332ff0e94fa" exitCode=0 Mar 21 05:08:06 crc kubenswrapper[4775]: I0321 05:08:06.249826 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64fb567758-hd2ld" event={"ID":"09cf3763-0b41-4452-a247-d9a56f58b05d","Type":"ContainerDied","Data":"1da6670625f81794dc89cffb06ac6d008a9c7c8ed1b16d49c09b2332ff0e94fa"} Mar 21 05:08:06 crc kubenswrapper[4775]: I0321 05:08:06.256364 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d02bb319-292e-449c-8e5b-42c6859f1529","Type":"ContainerStarted","Data":"617448a041c2ffe276be923a8975488baf07292190b1d4d69545558da5c6c9c9"} Mar 21 05:08:06 crc kubenswrapper[4775]: I0321 05:08:06.743210 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567828-6klnl" Mar 21 05:08:06 crc kubenswrapper[4775]: I0321 05:08:06.829436 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gt4qg\" (UniqueName: \"kubernetes.io/projected/eccf102b-f91a-4251-b878-1c21eea92522-kube-api-access-gt4qg\") pod \"eccf102b-f91a-4251-b878-1c21eea92522\" (UID: \"eccf102b-f91a-4251-b878-1c21eea92522\") " Mar 21 05:08:06 crc kubenswrapper[4775]: I0321 05:08:06.843435 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:08:06 crc kubenswrapper[4775]: I0321 05:08:06.875273 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eccf102b-f91a-4251-b878-1c21eea92522-kube-api-access-gt4qg" (OuterVolumeSpecName: "kube-api-access-gt4qg") pod "eccf102b-f91a-4251-b878-1c21eea92522" (UID: "eccf102b-f91a-4251-b878-1c21eea92522"). InnerVolumeSpecName "kube-api-access-gt4qg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:08:06 crc kubenswrapper[4775]: I0321 05:08:06.919342 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-bvxtx"] Mar 21 05:08:06 crc kubenswrapper[4775]: I0321 05:08:06.919644 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" podUID="f65ac8a3-4328-41af-8854-e8a2a9fce295" containerName="dnsmasq-dns" containerID="cri-o://18c26c2a6ec8c71c7881c713a6d2fb65eaa2258d47fd0f272a51e57881466532" gracePeriod=10 Mar 21 05:08:06 crc kubenswrapper[4775]: I0321 05:08:06.934503 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gt4qg\" (UniqueName: \"kubernetes.io/projected/eccf102b-f91a-4251-b878-1c21eea92522-kube-api-access-gt4qg\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.290088 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d02bb319-292e-449c-8e5b-42c6859f1529","Type":"ContainerStarted","Data":"f1a65482fa1320beaf95005d562b733e9a1acd6ed031fa590fa9340cdc95cc53"} Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.294560 4775 generic.go:334] "Generic (PLEG): container finished" podID="f65ac8a3-4328-41af-8854-e8a2a9fce295" containerID="18c26c2a6ec8c71c7881c713a6d2fb65eaa2258d47fd0f272a51e57881466532" exitCode=0 Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.294634 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" event={"ID":"f65ac8a3-4328-41af-8854-e8a2a9fce295","Type":"ContainerDied","Data":"18c26c2a6ec8c71c7881c713a6d2fb65eaa2258d47fd0f272a51e57881466532"} Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.305468 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567828-6klnl" event={"ID":"eccf102b-f91a-4251-b878-1c21eea92522","Type":"ContainerDied","Data":"486b3020022220622c2fb587ef965adb79954d3ca5c1cadb0f004636ab8a7a5f"} Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.305553 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="486b3020022220622c2fb587ef965adb79954d3ca5c1cadb0f004636ab8a7a5f" Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.305626 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567828-6klnl" Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.355199 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567822-pqlrw"] Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.386395 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567822-pqlrw"] Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.687502 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.690032 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f791f730-ffa2-44d3-b151-dc9d6e6e1743" path="/var/lib/kubelet/pods/f791f730-ffa2-44d3-b151-dc9d6e6e1743/volumes" Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.772611 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-dns-svc\") pod \"f65ac8a3-4328-41af-8854-e8a2a9fce295\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.772834 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-ovsdbserver-sb\") pod \"f65ac8a3-4328-41af-8854-e8a2a9fce295\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.772911 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vtjl\" (UniqueName: \"kubernetes.io/projected/f65ac8a3-4328-41af-8854-e8a2a9fce295-kube-api-access-6vtjl\") pod \"f65ac8a3-4328-41af-8854-e8a2a9fce295\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.772930 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-dns-swift-storage-0\") pod \"f65ac8a3-4328-41af-8854-e8a2a9fce295\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.772962 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-config\") pod \"f65ac8a3-4328-41af-8854-e8a2a9fce295\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.773126 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-ovsdbserver-nb\") pod \"f65ac8a3-4328-41af-8854-e8a2a9fce295\" (UID: \"f65ac8a3-4328-41af-8854-e8a2a9fce295\") " Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.783323 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f65ac8a3-4328-41af-8854-e8a2a9fce295-kube-api-access-6vtjl" (OuterVolumeSpecName: "kube-api-access-6vtjl") pod "f65ac8a3-4328-41af-8854-e8a2a9fce295" (UID: "f65ac8a3-4328-41af-8854-e8a2a9fce295"). InnerVolumeSpecName "kube-api-access-6vtjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.838844 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f65ac8a3-4328-41af-8854-e8a2a9fce295" (UID: "f65ac8a3-4328-41af-8854-e8a2a9fce295"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.857269 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f65ac8a3-4328-41af-8854-e8a2a9fce295" (UID: "f65ac8a3-4328-41af-8854-e8a2a9fce295"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.871657 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-config" (OuterVolumeSpecName: "config") pod "f65ac8a3-4328-41af-8854-e8a2a9fce295" (UID: "f65ac8a3-4328-41af-8854-e8a2a9fce295"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.875345 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.875377 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vtjl\" (UniqueName: \"kubernetes.io/projected/f65ac8a3-4328-41af-8854-e8a2a9fce295-kube-api-access-6vtjl\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.875389 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.875397 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.914315 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f65ac8a3-4328-41af-8854-e8a2a9fce295" (UID: "f65ac8a3-4328-41af-8854-e8a2a9fce295"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.948348 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f65ac8a3-4328-41af-8854-e8a2a9fce295" (UID: "f65ac8a3-4328-41af-8854-e8a2a9fce295"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.976537 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:07 crc kubenswrapper[4775]: I0321 05:08:07.976572 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f65ac8a3-4328-41af-8854-e8a2a9fce295-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.145180 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-66c879cfdd-smnxp" Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.319342 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d02bb319-292e-449c-8e5b-42c6859f1529","Type":"ContainerStarted","Data":"7218f41fdb3215097835d065eddc8dc9dc7b371f024366fe1ef833bd0e95e62d"} Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.326652 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" event={"ID":"f65ac8a3-4328-41af-8854-e8a2a9fce295","Type":"ContainerDied","Data":"b2b7a3537eba6e1449d61c2042541c00aac4345e771c671fe8c458e056f81a53"} Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.326693 4775 scope.go:117] "RemoveContainer" containerID="18c26c2a6ec8c71c7881c713a6d2fb65eaa2258d47fd0f272a51e57881466532" Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.326799 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-bvxtx" Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.343875 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.343857774 podStartE2EDuration="3.343857774s" podCreationTimestamp="2026-03-21 05:08:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:08:08.333753198 +0000 UTC m=+1241.310216832" watchObservedRunningTime="2026-03-21 05:08:08.343857774 +0000 UTC m=+1241.320321398" Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.359398 4775 scope.go:117] "RemoveContainer" containerID="73687f66a6e199eea3343a248d8eb3f006e4d6f0bdcb64641db9e3b7a2d85758" Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.371647 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-bvxtx"] Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.413856 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-bvxtx"] Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.859316 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 21 05:08:08 crc kubenswrapper[4775]: E0321 05:08:08.859953 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f65ac8a3-4328-41af-8854-e8a2a9fce295" containerName="init" Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.859976 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f65ac8a3-4328-41af-8854-e8a2a9fce295" containerName="init" Mar 21 05:08:08 crc kubenswrapper[4775]: E0321 05:08:08.859988 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f65ac8a3-4328-41af-8854-e8a2a9fce295" containerName="dnsmasq-dns" Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.859997 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f65ac8a3-4328-41af-8854-e8a2a9fce295" containerName="dnsmasq-dns" Mar 21 05:08:08 crc kubenswrapper[4775]: E0321 05:08:08.860041 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eccf102b-f91a-4251-b878-1c21eea92522" containerName="oc" Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.860050 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="eccf102b-f91a-4251-b878-1c21eea92522" containerName="oc" Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.860280 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="eccf102b-f91a-4251-b878-1c21eea92522" containerName="oc" Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.860305 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f65ac8a3-4328-41af-8854-e8a2a9fce295" containerName="dnsmasq-dns" Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.860993 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.864184 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-6nwpr" Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.864206 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.878741 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.882904 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 21 05:08:08 crc kubenswrapper[4775]: I0321 05:08:08.932553 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.009099 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dfe744b5-63ab-4df2-89cd-05605cfa4111-openstack-config\") pod \"openstackclient\" (UID: \"dfe744b5-63ab-4df2-89cd-05605cfa4111\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.009194 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbnqx\" (UniqueName: \"kubernetes.io/projected/dfe744b5-63ab-4df2-89cd-05605cfa4111-kube-api-access-bbnqx\") pod \"openstackclient\" (UID: \"dfe744b5-63ab-4df2-89cd-05605cfa4111\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.009252 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dfe744b5-63ab-4df2-89cd-05605cfa4111-openstack-config-secret\") pod \"openstackclient\" (UID: \"dfe744b5-63ab-4df2-89cd-05605cfa4111\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.009328 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfe744b5-63ab-4df2-89cd-05605cfa4111-combined-ca-bundle\") pod \"openstackclient\" (UID: \"dfe744b5-63ab-4df2-89cd-05605cfa4111\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.111399 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfe744b5-63ab-4df2-89cd-05605cfa4111-combined-ca-bundle\") pod \"openstackclient\" (UID: \"dfe744b5-63ab-4df2-89cd-05605cfa4111\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.111803 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dfe744b5-63ab-4df2-89cd-05605cfa4111-openstack-config\") pod \"openstackclient\" (UID: \"dfe744b5-63ab-4df2-89cd-05605cfa4111\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.111917 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbnqx\" (UniqueName: \"kubernetes.io/projected/dfe744b5-63ab-4df2-89cd-05605cfa4111-kube-api-access-bbnqx\") pod \"openstackclient\" (UID: \"dfe744b5-63ab-4df2-89cd-05605cfa4111\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.112010 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dfe744b5-63ab-4df2-89cd-05605cfa4111-openstack-config-secret\") pod \"openstackclient\" (UID: \"dfe744b5-63ab-4df2-89cd-05605cfa4111\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.112784 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dfe744b5-63ab-4df2-89cd-05605cfa4111-openstack-config\") pod \"openstackclient\" (UID: \"dfe744b5-63ab-4df2-89cd-05605cfa4111\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.117719 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dfe744b5-63ab-4df2-89cd-05605cfa4111-openstack-config-secret\") pod \"openstackclient\" (UID: \"dfe744b5-63ab-4df2-89cd-05605cfa4111\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.118235 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfe744b5-63ab-4df2-89cd-05605cfa4111-combined-ca-bundle\") pod \"openstackclient\" (UID: \"dfe744b5-63ab-4df2-89cd-05605cfa4111\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.127169 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbnqx\" (UniqueName: \"kubernetes.io/projected/dfe744b5-63ab-4df2-89cd-05605cfa4111-kube-api-access-bbnqx\") pod \"openstackclient\" (UID: \"dfe744b5-63ab-4df2-89cd-05605cfa4111\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.160733 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.161384 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.172675 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.216943 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.221417 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.234865 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.321159 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb5a0456-b5c5-433a-afde-fe38740e2310-combined-ca-bundle\") pod \"openstackclient\" (UID: \"bb5a0456-b5c5-433a-afde-fe38740e2310\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.321564 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bb5a0456-b5c5-433a-afde-fe38740e2310-openstack-config-secret\") pod \"openstackclient\" (UID: \"bb5a0456-b5c5-433a-afde-fe38740e2310\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.321609 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bb5a0456-b5c5-433a-afde-fe38740e2310-openstack-config\") pod \"openstackclient\" (UID: \"bb5a0456-b5c5-433a-afde-fe38740e2310\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.321661 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knd2b\" (UniqueName: \"kubernetes.io/projected/bb5a0456-b5c5-433a-afde-fe38740e2310-kube-api-access-knd2b\") pod \"openstackclient\" (UID: \"bb5a0456-b5c5-433a-afde-fe38740e2310\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: E0321 05:08:09.369985 4775 log.go:32] "RunPodSandbox from runtime service failed" err=< Mar 21 05:08:09 crc kubenswrapper[4775]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_dfe744b5-63ab-4df2-89cd-05605cfa4111_0(f8365ba4a6b3a7f175026f4a4a5c25207ab22c926e92448937608c75f2fd4f1b): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f8365ba4a6b3a7f175026f4a4a5c25207ab22c926e92448937608c75f2fd4f1b" Netns:"/var/run/netns/7b7609de-10ad-42d6-b5d7-23f5745f15d5" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=f8365ba4a6b3a7f175026f4a4a5c25207ab22c926e92448937608c75f2fd4f1b;K8S_POD_UID=dfe744b5-63ab-4df2-89cd-05605cfa4111" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/dfe744b5-63ab-4df2-89cd-05605cfa4111]: expected pod UID "dfe744b5-63ab-4df2-89cd-05605cfa4111" but got "bb5a0456-b5c5-433a-afde-fe38740e2310" from Kube API Mar 21 05:08:09 crc kubenswrapper[4775]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 21 05:08:09 crc kubenswrapper[4775]: > Mar 21 05:08:09 crc kubenswrapper[4775]: E0321 05:08:09.370069 4775 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Mar 21 05:08:09 crc kubenswrapper[4775]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_dfe744b5-63ab-4df2-89cd-05605cfa4111_0(f8365ba4a6b3a7f175026f4a4a5c25207ab22c926e92448937608c75f2fd4f1b): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f8365ba4a6b3a7f175026f4a4a5c25207ab22c926e92448937608c75f2fd4f1b" Netns:"/var/run/netns/7b7609de-10ad-42d6-b5d7-23f5745f15d5" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=f8365ba4a6b3a7f175026f4a4a5c25207ab22c926e92448937608c75f2fd4f1b;K8S_POD_UID=dfe744b5-63ab-4df2-89cd-05605cfa4111" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/dfe744b5-63ab-4df2-89cd-05605cfa4111]: expected pod UID "dfe744b5-63ab-4df2-89cd-05605cfa4111" but got "bb5a0456-b5c5-433a-afde-fe38740e2310" from Kube API Mar 21 05:08:09 crc kubenswrapper[4775]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 21 05:08:09 crc kubenswrapper[4775]: > pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.423535 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bb5a0456-b5c5-433a-afde-fe38740e2310-openstack-config-secret\") pod \"openstackclient\" (UID: \"bb5a0456-b5c5-433a-afde-fe38740e2310\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.423587 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bb5a0456-b5c5-433a-afde-fe38740e2310-openstack-config\") pod \"openstackclient\" (UID: \"bb5a0456-b5c5-433a-afde-fe38740e2310\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.423628 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knd2b\" (UniqueName: \"kubernetes.io/projected/bb5a0456-b5c5-433a-afde-fe38740e2310-kube-api-access-knd2b\") pod \"openstackclient\" (UID: \"bb5a0456-b5c5-433a-afde-fe38740e2310\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.423731 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb5a0456-b5c5-433a-afde-fe38740e2310-combined-ca-bundle\") pod \"openstackclient\" (UID: \"bb5a0456-b5c5-433a-afde-fe38740e2310\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.424646 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bb5a0456-b5c5-433a-afde-fe38740e2310-openstack-config\") pod \"openstackclient\" (UID: \"bb5a0456-b5c5-433a-afde-fe38740e2310\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.430182 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb5a0456-b5c5-433a-afde-fe38740e2310-combined-ca-bundle\") pod \"openstackclient\" (UID: \"bb5a0456-b5c5-433a-afde-fe38740e2310\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.430212 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bb5a0456-b5c5-433a-afde-fe38740e2310-openstack-config-secret\") pod \"openstackclient\" (UID: \"bb5a0456-b5c5-433a-afde-fe38740e2310\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.442712 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knd2b\" (UniqueName: \"kubernetes.io/projected/bb5a0456-b5c5-433a-afde-fe38740e2310-kube-api-access-knd2b\") pod \"openstackclient\" (UID: \"bb5a0456-b5c5-433a-afde-fe38740e2310\") " pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.542841 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.661351 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 21 05:08:09 crc kubenswrapper[4775]: I0321 05:08:09.678230 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f65ac8a3-4328-41af-8854-e8a2a9fce295" path="/var/lib/kubelet/pods/f65ac8a3-4328-41af-8854-e8a2a9fce295/volumes" Mar 21 05:08:10 crc kubenswrapper[4775]: I0321 05:08:10.039073 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 21 05:08:10 crc kubenswrapper[4775]: W0321 05:08:10.045934 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb5a0456_b5c5_433a_afde_fe38740e2310.slice/crio-fa9018ccabaf8b89d5be7fae7d1f288a596ea8f63313f06d19d0fbac3f5ae1f1 WatchSource:0}: Error finding container fa9018ccabaf8b89d5be7fae7d1f288a596ea8f63313f06d19d0fbac3f5ae1f1: Status 404 returned error can't find the container with id fa9018ccabaf8b89d5be7fae7d1f288a596ea8f63313f06d19d0fbac3f5ae1f1 Mar 21 05:08:10 crc kubenswrapper[4775]: I0321 05:08:10.358594 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"bb5a0456-b5c5-433a-afde-fe38740e2310","Type":"ContainerStarted","Data":"fa9018ccabaf8b89d5be7fae7d1f288a596ea8f63313f06d19d0fbac3f5ae1f1"} Mar 21 05:08:10 crc kubenswrapper[4775]: I0321 05:08:10.358604 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 21 05:08:10 crc kubenswrapper[4775]: I0321 05:08:10.368890 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 21 05:08:10 crc kubenswrapper[4775]: I0321 05:08:10.373561 4775 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="dfe744b5-63ab-4df2-89cd-05605cfa4111" podUID="bb5a0456-b5c5-433a-afde-fe38740e2310" Mar 21 05:08:10 crc kubenswrapper[4775]: I0321 05:08:10.542615 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dfe744b5-63ab-4df2-89cd-05605cfa4111-openstack-config-secret\") pod \"dfe744b5-63ab-4df2-89cd-05605cfa4111\" (UID: \"dfe744b5-63ab-4df2-89cd-05605cfa4111\") " Mar 21 05:08:10 crc kubenswrapper[4775]: I0321 05:08:10.542795 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dfe744b5-63ab-4df2-89cd-05605cfa4111-openstack-config\") pod \"dfe744b5-63ab-4df2-89cd-05605cfa4111\" (UID: \"dfe744b5-63ab-4df2-89cd-05605cfa4111\") " Mar 21 05:08:10 crc kubenswrapper[4775]: I0321 05:08:10.542826 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfe744b5-63ab-4df2-89cd-05605cfa4111-combined-ca-bundle\") pod \"dfe744b5-63ab-4df2-89cd-05605cfa4111\" (UID: \"dfe744b5-63ab-4df2-89cd-05605cfa4111\") " Mar 21 05:08:10 crc kubenswrapper[4775]: I0321 05:08:10.542909 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbnqx\" (UniqueName: \"kubernetes.io/projected/dfe744b5-63ab-4df2-89cd-05605cfa4111-kube-api-access-bbnqx\") pod \"dfe744b5-63ab-4df2-89cd-05605cfa4111\" (UID: \"dfe744b5-63ab-4df2-89cd-05605cfa4111\") " Mar 21 05:08:10 crc kubenswrapper[4775]: I0321 05:08:10.543937 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfe744b5-63ab-4df2-89cd-05605cfa4111-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "dfe744b5-63ab-4df2-89cd-05605cfa4111" (UID: "dfe744b5-63ab-4df2-89cd-05605cfa4111"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:08:10 crc kubenswrapper[4775]: I0321 05:08:10.548330 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfe744b5-63ab-4df2-89cd-05605cfa4111-kube-api-access-bbnqx" (OuterVolumeSpecName: "kube-api-access-bbnqx") pod "dfe744b5-63ab-4df2-89cd-05605cfa4111" (UID: "dfe744b5-63ab-4df2-89cd-05605cfa4111"). InnerVolumeSpecName "kube-api-access-bbnqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:08:10 crc kubenswrapper[4775]: I0321 05:08:10.560938 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfe744b5-63ab-4df2-89cd-05605cfa4111-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dfe744b5-63ab-4df2-89cd-05605cfa4111" (UID: "dfe744b5-63ab-4df2-89cd-05605cfa4111"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:10 crc kubenswrapper[4775]: I0321 05:08:10.562224 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfe744b5-63ab-4df2-89cd-05605cfa4111-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "dfe744b5-63ab-4df2-89cd-05605cfa4111" (UID: "dfe744b5-63ab-4df2-89cd-05605cfa4111"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:10 crc kubenswrapper[4775]: I0321 05:08:10.644945 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfe744b5-63ab-4df2-89cd-05605cfa4111-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:10 crc kubenswrapper[4775]: I0321 05:08:10.644983 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbnqx\" (UniqueName: \"kubernetes.io/projected/dfe744b5-63ab-4df2-89cd-05605cfa4111-kube-api-access-bbnqx\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:10 crc kubenswrapper[4775]: I0321 05:08:10.644998 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dfe744b5-63ab-4df2-89cd-05605cfa4111-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:10 crc kubenswrapper[4775]: I0321 05:08:10.645009 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dfe744b5-63ab-4df2-89cd-05605cfa4111-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:10 crc kubenswrapper[4775]: I0321 05:08:10.681517 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 21 05:08:11 crc kubenswrapper[4775]: I0321 05:08:11.374537 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 21 05:08:11 crc kubenswrapper[4775]: I0321 05:08:11.412953 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 21 05:08:11 crc kubenswrapper[4775]: I0321 05:08:11.448637 4775 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="dfe744b5-63ab-4df2-89cd-05605cfa4111" podUID="bb5a0456-b5c5-433a-afde-fe38740e2310" Mar 21 05:08:11 crc kubenswrapper[4775]: I0321 05:08:11.686093 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfe744b5-63ab-4df2-89cd-05605cfa4111" path="/var/lib/kubelet/pods/dfe744b5-63ab-4df2-89cd-05605cfa4111/volumes" Mar 21 05:08:12 crc kubenswrapper[4775]: I0321 05:08:12.538935 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:12 crc kubenswrapper[4775]: I0321 05:08:12.843801 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-77fd86567d-mf2wb" Mar 21 05:08:12 crc kubenswrapper[4775]: I0321 05:08:12.913251 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7dc8b6bf8b-btcqr"] Mar 21 05:08:12 crc kubenswrapper[4775]: I0321 05:08:12.913702 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" podUID="86abc8a1-1d5e-4dad-9e62-4f1c88608b95" containerName="barbican-api-log" containerID="cri-o://da40033a5dd52e534fbf0e897d41a379c55c497ad81da6c097f013d7acf158d3" gracePeriod=30 Mar 21 05:08:12 crc kubenswrapper[4775]: I0321 05:08:12.914126 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" podUID="86abc8a1-1d5e-4dad-9e62-4f1c88608b95" containerName="barbican-api" containerID="cri-o://9e09a5090c4d61aabf6c91b2097f27d3892ae005a5644d72b8acf7ad49174027" gracePeriod=30 Mar 21 05:08:12 crc kubenswrapper[4775]: I0321 05:08:12.926642 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" podUID="86abc8a1-1d5e-4dad-9e62-4f1c88608b95" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.166:9311/healthcheck\": EOF" Mar 21 05:08:13 crc kubenswrapper[4775]: I0321 05:08:13.422988 4775 generic.go:334] "Generic (PLEG): container finished" podID="86abc8a1-1d5e-4dad-9e62-4f1c88608b95" containerID="da40033a5dd52e534fbf0e897d41a379c55c497ad81da6c097f013d7acf158d3" exitCode=143 Mar 21 05:08:13 crc kubenswrapper[4775]: I0321 05:08:13.423043 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" event={"ID":"86abc8a1-1d5e-4dad-9e62-4f1c88608b95","Type":"ContainerDied","Data":"da40033a5dd52e534fbf0e897d41a379c55c497ad81da6c097f013d7acf158d3"} Mar 21 05:08:14 crc kubenswrapper[4775]: I0321 05:08:14.951369 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-f9b88fb79-vclnv"] Mar 21 05:08:14 crc kubenswrapper[4775]: I0321 05:08:14.957829 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:14 crc kubenswrapper[4775]: I0321 05:08:14.960540 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 21 05:08:14 crc kubenswrapper[4775]: I0321 05:08:14.960565 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-f9b88fb79-vclnv"] Mar 21 05:08:14 crc kubenswrapper[4775]: I0321 05:08:14.960703 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Mar 21 05:08:14 crc kubenswrapper[4775]: I0321 05:08:14.960706 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Mar 21 05:08:14 crc kubenswrapper[4775]: I0321 05:08:14.979464 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-64fb567758-hd2ld" podUID="09cf3763-0b41-4452-a247-d9a56f58b05d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.082772 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfjng\" (UniqueName: \"kubernetes.io/projected/77432545-f22c-453a-b6a7-7c932712efa9-kube-api-access-mfjng\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.082823 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/77432545-f22c-453a-b6a7-7c932712efa9-etc-swift\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.082914 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77432545-f22c-453a-b6a7-7c932712efa9-combined-ca-bundle\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.083000 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77432545-f22c-453a-b6a7-7c932712efa9-public-tls-certs\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.083076 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77432545-f22c-453a-b6a7-7c932712efa9-config-data\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.083361 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77432545-f22c-453a-b6a7-7c932712efa9-run-httpd\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.083454 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77432545-f22c-453a-b6a7-7c932712efa9-internal-tls-certs\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.083527 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77432545-f22c-453a-b6a7-7c932712efa9-log-httpd\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.184995 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfjng\" (UniqueName: \"kubernetes.io/projected/77432545-f22c-453a-b6a7-7c932712efa9-kube-api-access-mfjng\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.185046 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/77432545-f22c-453a-b6a7-7c932712efa9-etc-swift\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.185103 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77432545-f22c-453a-b6a7-7c932712efa9-combined-ca-bundle\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.185144 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77432545-f22c-453a-b6a7-7c932712efa9-public-tls-certs\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.185165 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77432545-f22c-453a-b6a7-7c932712efa9-config-data\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.185224 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77432545-f22c-453a-b6a7-7c932712efa9-run-httpd\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.185255 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77432545-f22c-453a-b6a7-7c932712efa9-internal-tls-certs\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.185277 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77432545-f22c-453a-b6a7-7c932712efa9-log-httpd\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.185760 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77432545-f22c-453a-b6a7-7c932712efa9-log-httpd\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.187937 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77432545-f22c-453a-b6a7-7c932712efa9-run-httpd\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.192182 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77432545-f22c-453a-b6a7-7c932712efa9-combined-ca-bundle\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.193019 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77432545-f22c-453a-b6a7-7c932712efa9-config-data\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.192982 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77432545-f22c-453a-b6a7-7c932712efa9-internal-tls-certs\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.194084 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/77432545-f22c-453a-b6a7-7c932712efa9-etc-swift\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.208915 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77432545-f22c-453a-b6a7-7c932712efa9-public-tls-certs\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.210978 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfjng\" (UniqueName: \"kubernetes.io/projected/77432545-f22c-453a-b6a7-7c932712efa9-kube-api-access-mfjng\") pod \"swift-proxy-f9b88fb79-vclnv\" (UID: \"77432545-f22c-453a-b6a7-7c932712efa9\") " pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.295350 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:15 crc kubenswrapper[4775]: I0321 05:08:15.945341 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-f9b88fb79-vclnv"] Mar 21 05:08:15 crc kubenswrapper[4775]: W0321 05:08:15.958393 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77432545_f22c_453a_b6a7_7c932712efa9.slice/crio-bcdf50870f46b4a992b7a1625618f6249d5ebc5618940c1849f89429598adbec WatchSource:0}: Error finding container bcdf50870f46b4a992b7a1625618f6249d5ebc5618940c1849f89429598adbec: Status 404 returned error can't find the container with id bcdf50870f46b4a992b7a1625618f6249d5ebc5618940c1849f89429598adbec Mar 21 05:08:16 crc kubenswrapper[4775]: I0321 05:08:16.283025 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 21 05:08:16 crc kubenswrapper[4775]: I0321 05:08:16.472902 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-f9b88fb79-vclnv" event={"ID":"77432545-f22c-453a-b6a7-7c932712efa9","Type":"ContainerStarted","Data":"bcdf50870f46b4a992b7a1625618f6249d5ebc5618940c1849f89429598adbec"} Mar 21 05:08:16 crc kubenswrapper[4775]: I0321 05:08:16.483669 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="8ac5c16c-56ac-4299-ae61-a8200986ce10" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 21 05:08:16 crc kubenswrapper[4775]: I0321 05:08:16.976522 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 21 05:08:16 crc kubenswrapper[4775]: I0321 05:08:16.990440 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.182520 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-tmm94"] Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.189407 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-tmm94" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.194043 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-tmm94"] Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.328263 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc238c51-34cc-4cd6-bdf2-00e3747a9d67-operator-scripts\") pod \"nova-api-db-create-tmm94\" (UID: \"bc238c51-34cc-4cd6-bdf2-00e3747a9d67\") " pod="openstack/nova-api-db-create-tmm94" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.328398 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkh9x\" (UniqueName: \"kubernetes.io/projected/bc238c51-34cc-4cd6-bdf2-00e3747a9d67-kube-api-access-xkh9x\") pod \"nova-api-db-create-tmm94\" (UID: \"bc238c51-34cc-4cd6-bdf2-00e3747a9d67\") " pod="openstack/nova-api-db-create-tmm94" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.382264 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-wxmkg"] Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.383753 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wxmkg" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.390172 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-a38c-account-create-update-x226p"] Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.391622 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a38c-account-create-update-x226p" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.394557 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.397878 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-wxmkg"] Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.405219 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-a38c-account-create-update-x226p"] Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.418558 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" podUID="86abc8a1-1d5e-4dad-9e62-4f1c88608b95" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.166:9311/healthcheck\": read tcp 10.217.0.2:58266->10.217.0.166:9311: read: connection reset by peer" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.418597 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" podUID="86abc8a1-1d5e-4dad-9e62-4f1c88608b95" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.166:9311/healthcheck\": read tcp 10.217.0.2:58256->10.217.0.166:9311: read: connection reset by peer" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.432510 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b65ed673-344a-4064-ae90-c9d20964c648-operator-scripts\") pod \"nova-cell0-db-create-wxmkg\" (UID: \"b65ed673-344a-4064-ae90-c9d20964c648\") " pod="openstack/nova-cell0-db-create-wxmkg" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.432570 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc238c51-34cc-4cd6-bdf2-00e3747a9d67-operator-scripts\") pod \"nova-api-db-create-tmm94\" (UID: \"bc238c51-34cc-4cd6-bdf2-00e3747a9d67\") " pod="openstack/nova-api-db-create-tmm94" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.432600 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb6wb\" (UniqueName: \"kubernetes.io/projected/b65ed673-344a-4064-ae90-c9d20964c648-kube-api-access-wb6wb\") pod \"nova-cell0-db-create-wxmkg\" (UID: \"b65ed673-344a-4064-ae90-c9d20964c648\") " pod="openstack/nova-cell0-db-create-wxmkg" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.432709 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkh9x\" (UniqueName: \"kubernetes.io/projected/bc238c51-34cc-4cd6-bdf2-00e3747a9d67-kube-api-access-xkh9x\") pod \"nova-api-db-create-tmm94\" (UID: \"bc238c51-34cc-4cd6-bdf2-00e3747a9d67\") " pod="openstack/nova-api-db-create-tmm94" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.433916 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc238c51-34cc-4cd6-bdf2-00e3747a9d67-operator-scripts\") pod \"nova-api-db-create-tmm94\" (UID: \"bc238c51-34cc-4cd6-bdf2-00e3747a9d67\") " pod="openstack/nova-api-db-create-tmm94" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.451575 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkh9x\" (UniqueName: \"kubernetes.io/projected/bc238c51-34cc-4cd6-bdf2-00e3747a9d67-kube-api-access-xkh9x\") pod \"nova-api-db-create-tmm94\" (UID: \"bc238c51-34cc-4cd6-bdf2-00e3747a9d67\") " pod="openstack/nova-api-db-create-tmm94" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.493247 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-zfz4w"] Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.494551 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-zfz4w" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.514407 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-f9b88fb79-vclnv" event={"ID":"77432545-f22c-453a-b6a7-7c932712efa9","Type":"ContainerStarted","Data":"9eedba89840a3fd35f983095f0bd1a2e6c833d019237a66f24e8a6d46f58b2aa"} Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.529345 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-zfz4w"] Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.529819 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-tmm94" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.531744 4775 generic.go:334] "Generic (PLEG): container finished" podID="86abc8a1-1d5e-4dad-9e62-4f1c88608b95" containerID="9e09a5090c4d61aabf6c91b2097f27d3892ae005a5644d72b8acf7ad49174027" exitCode=0 Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.532183 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" event={"ID":"86abc8a1-1d5e-4dad-9e62-4f1c88608b95","Type":"ContainerDied","Data":"9e09a5090c4d61aabf6c91b2097f27d3892ae005a5644d72b8acf7ad49174027"} Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.537583 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/028d46f7-6f14-40f6-a6cf-77aeec59a99e-operator-scripts\") pod \"nova-api-a38c-account-create-update-x226p\" (UID: \"028d46f7-6f14-40f6-a6cf-77aeec59a99e\") " pod="openstack/nova-api-a38c-account-create-update-x226p" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.539656 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcbnj\" (UniqueName: \"kubernetes.io/projected/028d46f7-6f14-40f6-a6cf-77aeec59a99e-kube-api-access-jcbnj\") pod \"nova-api-a38c-account-create-update-x226p\" (UID: \"028d46f7-6f14-40f6-a6cf-77aeec59a99e\") " pod="openstack/nova-api-a38c-account-create-update-x226p" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.539709 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b65ed673-344a-4064-ae90-c9d20964c648-operator-scripts\") pod \"nova-cell0-db-create-wxmkg\" (UID: \"b65ed673-344a-4064-ae90-c9d20964c648\") " pod="openstack/nova-cell0-db-create-wxmkg" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.539747 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb6wb\" (UniqueName: \"kubernetes.io/projected/b65ed673-344a-4064-ae90-c9d20964c648-kube-api-access-wb6wb\") pod \"nova-cell0-db-create-wxmkg\" (UID: \"b65ed673-344a-4064-ae90-c9d20964c648\") " pod="openstack/nova-cell0-db-create-wxmkg" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.540585 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b65ed673-344a-4064-ae90-c9d20964c648-operator-scripts\") pod \"nova-cell0-db-create-wxmkg\" (UID: \"b65ed673-344a-4064-ae90-c9d20964c648\") " pod="openstack/nova-cell0-db-create-wxmkg" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.558631 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb6wb\" (UniqueName: \"kubernetes.io/projected/b65ed673-344a-4064-ae90-c9d20964c648-kube-api-access-wb6wb\") pod \"nova-cell0-db-create-wxmkg\" (UID: \"b65ed673-344a-4064-ae90-c9d20964c648\") " pod="openstack/nova-cell0-db-create-wxmkg" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.586842 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db3c-account-create-update-k29zp"] Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.588853 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db3c-account-create-update-k29zp" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.591896 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.612581 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db3c-account-create-update-k29zp"] Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.641948 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87e82da2-fdd4-4ead-8a9b-022e55a03690-operator-scripts\") pod \"nova-cell1-db-create-zfz4w\" (UID: \"87e82da2-fdd4-4ead-8a9b-022e55a03690\") " pod="openstack/nova-cell1-db-create-zfz4w" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.642052 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/028d46f7-6f14-40f6-a6cf-77aeec59a99e-operator-scripts\") pod \"nova-api-a38c-account-create-update-x226p\" (UID: \"028d46f7-6f14-40f6-a6cf-77aeec59a99e\") " pod="openstack/nova-api-a38c-account-create-update-x226p" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.642219 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22a5e145-5b67-4d02-9104-ade3e48888db-operator-scripts\") pod \"nova-cell0-db3c-account-create-update-k29zp\" (UID: \"22a5e145-5b67-4d02-9104-ade3e48888db\") " pod="openstack/nova-cell0-db3c-account-create-update-k29zp" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.642250 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqfwp\" (UniqueName: \"kubernetes.io/projected/22a5e145-5b67-4d02-9104-ade3e48888db-kube-api-access-lqfwp\") pod \"nova-cell0-db3c-account-create-update-k29zp\" (UID: \"22a5e145-5b67-4d02-9104-ade3e48888db\") " pod="openstack/nova-cell0-db3c-account-create-update-k29zp" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.642313 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txzb7\" (UniqueName: \"kubernetes.io/projected/87e82da2-fdd4-4ead-8a9b-022e55a03690-kube-api-access-txzb7\") pod \"nova-cell1-db-create-zfz4w\" (UID: \"87e82da2-fdd4-4ead-8a9b-022e55a03690\") " pod="openstack/nova-cell1-db-create-zfz4w" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.642529 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcbnj\" (UniqueName: \"kubernetes.io/projected/028d46f7-6f14-40f6-a6cf-77aeec59a99e-kube-api-access-jcbnj\") pod \"nova-api-a38c-account-create-update-x226p\" (UID: \"028d46f7-6f14-40f6-a6cf-77aeec59a99e\") " pod="openstack/nova-api-a38c-account-create-update-x226p" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.643173 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/028d46f7-6f14-40f6-a6cf-77aeec59a99e-operator-scripts\") pod \"nova-api-a38c-account-create-update-x226p\" (UID: \"028d46f7-6f14-40f6-a6cf-77aeec59a99e\") " pod="openstack/nova-api-a38c-account-create-update-x226p" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.666500 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcbnj\" (UniqueName: \"kubernetes.io/projected/028d46f7-6f14-40f6-a6cf-77aeec59a99e-kube-api-access-jcbnj\") pod \"nova-api-a38c-account-create-update-x226p\" (UID: \"028d46f7-6f14-40f6-a6cf-77aeec59a99e\") " pod="openstack/nova-api-a38c-account-create-update-x226p" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.699264 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wxmkg" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.723945 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a38c-account-create-update-x226p" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.744472 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22a5e145-5b67-4d02-9104-ade3e48888db-operator-scripts\") pod \"nova-cell0-db3c-account-create-update-k29zp\" (UID: \"22a5e145-5b67-4d02-9104-ade3e48888db\") " pod="openstack/nova-cell0-db3c-account-create-update-k29zp" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.744528 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqfwp\" (UniqueName: \"kubernetes.io/projected/22a5e145-5b67-4d02-9104-ade3e48888db-kube-api-access-lqfwp\") pod \"nova-cell0-db3c-account-create-update-k29zp\" (UID: \"22a5e145-5b67-4d02-9104-ade3e48888db\") " pod="openstack/nova-cell0-db3c-account-create-update-k29zp" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.744558 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txzb7\" (UniqueName: \"kubernetes.io/projected/87e82da2-fdd4-4ead-8a9b-022e55a03690-kube-api-access-txzb7\") pod \"nova-cell1-db-create-zfz4w\" (UID: \"87e82da2-fdd4-4ead-8a9b-022e55a03690\") " pod="openstack/nova-cell1-db-create-zfz4w" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.744703 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87e82da2-fdd4-4ead-8a9b-022e55a03690-operator-scripts\") pod \"nova-cell1-db-create-zfz4w\" (UID: \"87e82da2-fdd4-4ead-8a9b-022e55a03690\") " pod="openstack/nova-cell1-db-create-zfz4w" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.745413 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22a5e145-5b67-4d02-9104-ade3e48888db-operator-scripts\") pod \"nova-cell0-db3c-account-create-update-k29zp\" (UID: \"22a5e145-5b67-4d02-9104-ade3e48888db\") " pod="openstack/nova-cell0-db3c-account-create-update-k29zp" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.745467 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87e82da2-fdd4-4ead-8a9b-022e55a03690-operator-scripts\") pod \"nova-cell1-db-create-zfz4w\" (UID: \"87e82da2-fdd4-4ead-8a9b-022e55a03690\") " pod="openstack/nova-cell1-db-create-zfz4w" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.770890 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txzb7\" (UniqueName: \"kubernetes.io/projected/87e82da2-fdd4-4ead-8a9b-022e55a03690-kube-api-access-txzb7\") pod \"nova-cell1-db-create-zfz4w\" (UID: \"87e82da2-fdd4-4ead-8a9b-022e55a03690\") " pod="openstack/nova-cell1-db-create-zfz4w" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.775005 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqfwp\" (UniqueName: \"kubernetes.io/projected/22a5e145-5b67-4d02-9104-ade3e48888db-kube-api-access-lqfwp\") pod \"nova-cell0-db3c-account-create-update-k29zp\" (UID: \"22a5e145-5b67-4d02-9104-ade3e48888db\") " pod="openstack/nova-cell0-db3c-account-create-update-k29zp" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.805442 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-45fb-account-create-update-glct2"] Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.807241 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-45fb-account-create-update-glct2" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.810920 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.827254 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-45fb-account-create-update-glct2"] Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.836530 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-zfz4w" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.921733 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db3c-account-create-update-k29zp" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.949501 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/330e8d60-7af3-42ed-a0d7-234264037d09-operator-scripts\") pod \"nova-cell1-45fb-account-create-update-glct2\" (UID: \"330e8d60-7af3-42ed-a0d7-234264037d09\") " pod="openstack/nova-cell1-45fb-account-create-update-glct2" Mar 21 05:08:17 crc kubenswrapper[4775]: I0321 05:08:17.949570 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47c7f\" (UniqueName: \"kubernetes.io/projected/330e8d60-7af3-42ed-a0d7-234264037d09-kube-api-access-47c7f\") pod \"nova-cell1-45fb-account-create-update-glct2\" (UID: \"330e8d60-7af3-42ed-a0d7-234264037d09\") " pod="openstack/nova-cell1-45fb-account-create-update-glct2" Mar 21 05:08:18 crc kubenswrapper[4775]: I0321 05:08:18.051954 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/330e8d60-7af3-42ed-a0d7-234264037d09-operator-scripts\") pod \"nova-cell1-45fb-account-create-update-glct2\" (UID: \"330e8d60-7af3-42ed-a0d7-234264037d09\") " pod="openstack/nova-cell1-45fb-account-create-update-glct2" Mar 21 05:08:18 crc kubenswrapper[4775]: I0321 05:08:18.052014 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47c7f\" (UniqueName: \"kubernetes.io/projected/330e8d60-7af3-42ed-a0d7-234264037d09-kube-api-access-47c7f\") pod \"nova-cell1-45fb-account-create-update-glct2\" (UID: \"330e8d60-7af3-42ed-a0d7-234264037d09\") " pod="openstack/nova-cell1-45fb-account-create-update-glct2" Mar 21 05:08:18 crc kubenswrapper[4775]: I0321 05:08:18.052698 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/330e8d60-7af3-42ed-a0d7-234264037d09-operator-scripts\") pod \"nova-cell1-45fb-account-create-update-glct2\" (UID: \"330e8d60-7af3-42ed-a0d7-234264037d09\") " pod="openstack/nova-cell1-45fb-account-create-update-glct2" Mar 21 05:08:18 crc kubenswrapper[4775]: I0321 05:08:18.070986 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47c7f\" (UniqueName: \"kubernetes.io/projected/330e8d60-7af3-42ed-a0d7-234264037d09-kube-api-access-47c7f\") pod \"nova-cell1-45fb-account-create-update-glct2\" (UID: \"330e8d60-7af3-42ed-a0d7-234264037d09\") " pod="openstack/nova-cell1-45fb-account-create-update-glct2" Mar 21 05:08:18 crc kubenswrapper[4775]: I0321 05:08:18.181757 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-45fb-account-create-update-glct2" Mar 21 05:08:19 crc kubenswrapper[4775]: I0321 05:08:19.500200 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:08:19 crc kubenswrapper[4775]: I0321 05:08:19.555680 4775 generic.go:334] "Generic (PLEG): container finished" podID="8ac5c16c-56ac-4299-ae61-a8200986ce10" containerID="507f523f41fa20fba933e438c92908136440aa69787cc4ade1ab3f5729729a60" exitCode=137 Mar 21 05:08:19 crc kubenswrapper[4775]: I0321 05:08:19.555733 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ac5c16c-56ac-4299-ae61-a8200986ce10","Type":"ContainerDied","Data":"507f523f41fa20fba933e438c92908136440aa69787cc4ade1ab3f5729729a60"} Mar 21 05:08:20 crc kubenswrapper[4775]: I0321 05:08:20.990444 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:08:20 crc kubenswrapper[4775]: I0321 05:08:20.993088 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-58948d8bb4-rcw89" Mar 21 05:08:21 crc kubenswrapper[4775]: I0321 05:08:21.852036 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" podUID="86abc8a1-1d5e-4dad-9e62-4f1c88608b95" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.166:9311/healthcheck\": dial tcp 10.217.0.166:9311: connect: connection refused" Mar 21 05:08:21 crc kubenswrapper[4775]: I0321 05:08:21.852053 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" podUID="86abc8a1-1d5e-4dad-9e62-4f1c88608b95" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.166:9311/healthcheck\": dial tcp 10.217.0.166:9311: connect: connection refused" Mar 21 05:08:24 crc kubenswrapper[4775]: I0321 05:08:24.980004 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-64fb567758-hd2ld" podUID="09cf3763-0b41-4452-a247-d9a56f58b05d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.478525 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.593711 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.651671 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ac5c16c-56ac-4299-ae61-a8200986ce10","Type":"ContainerDied","Data":"25ef4b17ffe6969bdd8510a4d651e67244e698c334397d84b7e113488d96c6fb"} Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.651714 4775 scope.go:117] "RemoveContainer" containerID="507f523f41fa20fba933e438c92908136440aa69787cc4ade1ab3f5729729a60" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.651823 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.661172 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.678460 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bzmg\" (UniqueName: \"kubernetes.io/projected/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-kube-api-access-6bzmg\") pod \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\" (UID: \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\") " Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.678735 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-config-data\") pod \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\" (UID: \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\") " Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.678813 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-combined-ca-bundle\") pod \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\" (UID: \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\") " Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.678921 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-logs\") pod \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\" (UID: \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\") " Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.679052 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-config-data-custom\") pod \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\" (UID: \"86abc8a1-1d5e-4dad-9e62-4f1c88608b95\") " Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.682596 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-logs" (OuterVolumeSpecName: "logs") pod "86abc8a1-1d5e-4dad-9e62-4f1c88608b95" (UID: "86abc8a1-1d5e-4dad-9e62-4f1c88608b95"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.693521 4775 scope.go:117] "RemoveContainer" containerID="3ffce781b6baa0a2f3ae718a0cd398633d05ff48719dba863adc5152b47fe849" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.696289 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "86abc8a1-1d5e-4dad-9e62-4f1c88608b95" (UID: "86abc8a1-1d5e-4dad-9e62-4f1c88608b95"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.701104 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-kube-api-access-6bzmg" (OuterVolumeSpecName: "kube-api-access-6bzmg") pod "86abc8a1-1d5e-4dad-9e62-4f1c88608b95" (UID: "86abc8a1-1d5e-4dad-9e62-4f1c88608b95"). InnerVolumeSpecName "kube-api-access-6bzmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.734526 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86abc8a1-1d5e-4dad-9e62-4f1c88608b95" (UID: "86abc8a1-1d5e-4dad-9e62-4f1c88608b95"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.742390 4775 scope.go:117] "RemoveContainer" containerID="4e1e7c557d6fa58625eeced787347009a65ed5b8a3da8a6487f343ffff4e6c90" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.759946 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7dc8b6bf8b-btcqr" event={"ID":"86abc8a1-1d5e-4dad-9e62-4f1c88608b95","Type":"ContainerDied","Data":"56894b1d2f57fb4749d6394cfec85a58eb1ab9e8c5e0205ea22bb5116fe8c5c9"} Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.778829 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-config-data" (OuterVolumeSpecName: "config-data") pod "86abc8a1-1d5e-4dad-9e62-4f1c88608b95" (UID: "86abc8a1-1d5e-4dad-9e62-4f1c88608b95"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.780524 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ac5c16c-56ac-4299-ae61-a8200986ce10-run-httpd\") pod \"8ac5c16c-56ac-4299-ae61-a8200986ce10\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.780590 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-scripts\") pod \"8ac5c16c-56ac-4299-ae61-a8200986ce10\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.780625 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-combined-ca-bundle\") pod \"8ac5c16c-56ac-4299-ae61-a8200986ce10\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.780756 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ac5c16c-56ac-4299-ae61-a8200986ce10-log-httpd\") pod \"8ac5c16c-56ac-4299-ae61-a8200986ce10\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.780916 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8h8k\" (UniqueName: \"kubernetes.io/projected/8ac5c16c-56ac-4299-ae61-a8200986ce10-kube-api-access-w8h8k\") pod \"8ac5c16c-56ac-4299-ae61-a8200986ce10\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.780979 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-config-data\") pod \"8ac5c16c-56ac-4299-ae61-a8200986ce10\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.781138 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-sg-core-conf-yaml\") pod \"8ac5c16c-56ac-4299-ae61-a8200986ce10\" (UID: \"8ac5c16c-56ac-4299-ae61-a8200986ce10\") " Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.781852 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bzmg\" (UniqueName: \"kubernetes.io/projected/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-kube-api-access-6bzmg\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.781892 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.781905 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.781918 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-logs\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.781930 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/86abc8a1-1d5e-4dad-9e62-4f1c88608b95-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.785969 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ac5c16c-56ac-4299-ae61-a8200986ce10-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8ac5c16c-56ac-4299-ae61-a8200986ce10" (UID: "8ac5c16c-56ac-4299-ae61-a8200986ce10"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.787359 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ac5c16c-56ac-4299-ae61-a8200986ce10-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8ac5c16c-56ac-4299-ae61-a8200986ce10" (UID: "8ac5c16c-56ac-4299-ae61-a8200986ce10"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.787850 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ac5c16c-56ac-4299-ae61-a8200986ce10-kube-api-access-w8h8k" (OuterVolumeSpecName: "kube-api-access-w8h8k") pod "8ac5c16c-56ac-4299-ae61-a8200986ce10" (UID: "8ac5c16c-56ac-4299-ae61-a8200986ce10"). InnerVolumeSpecName "kube-api-access-w8h8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.790439 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-scripts" (OuterVolumeSpecName: "scripts") pod "8ac5c16c-56ac-4299-ae61-a8200986ce10" (UID: "8ac5c16c-56ac-4299-ae61-a8200986ce10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.787602 4775 scope.go:117] "RemoveContainer" containerID="9e09a5090c4d61aabf6c91b2097f27d3892ae005a5644d72b8acf7ad49174027" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.817496 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8ac5c16c-56ac-4299-ae61-a8200986ce10" (UID: "8ac5c16c-56ac-4299-ae61-a8200986ce10"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.858613 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ac5c16c-56ac-4299-ae61-a8200986ce10" (UID: "8ac5c16c-56ac-4299-ae61-a8200986ce10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.869215 4775 scope.go:117] "RemoveContainer" containerID="da40033a5dd52e534fbf0e897d41a379c55c497ad81da6c097f013d7acf158d3" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.884252 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ac5c16c-56ac-4299-ae61-a8200986ce10-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.884292 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8h8k\" (UniqueName: \"kubernetes.io/projected/8ac5c16c-56ac-4299-ae61-a8200986ce10-kube-api-access-w8h8k\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.884301 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.884309 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ac5c16c-56ac-4299-ae61-a8200986ce10-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.884317 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.884325 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.899727 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-config-data" (OuterVolumeSpecName: "config-data") pod "8ac5c16c-56ac-4299-ae61-a8200986ce10" (UID: "8ac5c16c-56ac-4299-ae61-a8200986ce10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.926552 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7d65998c7c-prp5b" Mar 21 05:08:25 crc kubenswrapper[4775]: I0321 05:08:25.986758 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ac5c16c-56ac-4299-ae61-a8200986ce10-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:26 crc kubenswrapper[4775]: W0321 05:08:26.017416 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22a5e145_5b67_4d02_9104_ade3e48888db.slice/crio-e609ddd387c7c54bd1869b55edbc31d421deac134c26c25fe6a20a50afdcea02 WatchSource:0}: Error finding container e609ddd387c7c54bd1869b55edbc31d421deac134c26c25fe6a20a50afdcea02: Status 404 returned error can't find the container with id e609ddd387c7c54bd1869b55edbc31d421deac134c26c25fe6a20a50afdcea02 Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.026997 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db3c-account-create-update-k29zp"] Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.043057 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-68ccf5bf68-lf5dz"] Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.043333 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-68ccf5bf68-lf5dz" podUID="6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4" containerName="neutron-api" containerID="cri-o://f14a2e346ace37be895dcc94dd25c65e41716e7e754b07d2e1e3277674aeac90" gracePeriod=30 Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.043477 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-68ccf5bf68-lf5dz" podUID="6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4" containerName="neutron-httpd" containerID="cri-o://06c961a967b1c07bf826487fd3c050eb92ff3106f515b9136df5283020977b92" gracePeriod=30 Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.098849 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.125216 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.142296 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7dc8b6bf8b-btcqr"] Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.152077 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7dc8b6bf8b-btcqr"] Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.164620 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:08:26 crc kubenswrapper[4775]: E0321 05:08:26.165079 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ac5c16c-56ac-4299-ae61-a8200986ce10" containerName="proxy-httpd" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.165095 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ac5c16c-56ac-4299-ae61-a8200986ce10" containerName="proxy-httpd" Mar 21 05:08:26 crc kubenswrapper[4775]: E0321 05:08:26.165134 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86abc8a1-1d5e-4dad-9e62-4f1c88608b95" containerName="barbican-api-log" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.165143 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="86abc8a1-1d5e-4dad-9e62-4f1c88608b95" containerName="barbican-api-log" Mar 21 05:08:26 crc kubenswrapper[4775]: E0321 05:08:26.165161 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86abc8a1-1d5e-4dad-9e62-4f1c88608b95" containerName="barbican-api" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.165169 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="86abc8a1-1d5e-4dad-9e62-4f1c88608b95" containerName="barbican-api" Mar 21 05:08:26 crc kubenswrapper[4775]: E0321 05:08:26.165195 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ac5c16c-56ac-4299-ae61-a8200986ce10" containerName="sg-core" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.165202 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ac5c16c-56ac-4299-ae61-a8200986ce10" containerName="sg-core" Mar 21 05:08:26 crc kubenswrapper[4775]: E0321 05:08:26.165214 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ac5c16c-56ac-4299-ae61-a8200986ce10" containerName="ceilometer-notification-agent" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.165221 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ac5c16c-56ac-4299-ae61-a8200986ce10" containerName="ceilometer-notification-agent" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.165437 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ac5c16c-56ac-4299-ae61-a8200986ce10" containerName="ceilometer-notification-agent" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.165452 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ac5c16c-56ac-4299-ae61-a8200986ce10" containerName="sg-core" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.165472 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="86abc8a1-1d5e-4dad-9e62-4f1c88608b95" containerName="barbican-api" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.165483 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="86abc8a1-1d5e-4dad-9e62-4f1c88608b95" containerName="barbican-api-log" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.165492 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ac5c16c-56ac-4299-ae61-a8200986ce10" containerName="proxy-httpd" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.173451 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.175388 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.178370 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.178859 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.295532 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-run-httpd\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.295900 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-config-data\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.295947 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-log-httpd\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.295974 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.296079 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.296188 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-scripts\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.296380 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cmhz\" (UniqueName: \"kubernetes.io/projected/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-kube-api-access-8cmhz\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.372555 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-45fb-account-create-update-glct2"] Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.398756 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-config-data\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.398828 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-log-httpd\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.398866 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.398986 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.399042 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-scripts\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.399096 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cmhz\" (UniqueName: \"kubernetes.io/projected/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-kube-api-access-8cmhz\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.399225 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-run-httpd\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.399744 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-run-httpd\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.401434 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-log-httpd\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.416797 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-wxmkg"] Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.425470 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-scripts\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.428064 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cmhz\" (UniqueName: \"kubernetes.io/projected/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-kube-api-access-8cmhz\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.434473 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-config-data\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.440235 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.446154 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.451305 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-a38c-account-create-update-x226p"] Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.468734 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-tmm94"] Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.478444 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-zfz4w"] Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.600602 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.690650 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wxmkg" event={"ID":"b65ed673-344a-4064-ae90-c9d20964c648","Type":"ContainerStarted","Data":"a48ddd114a87c3d3cbfffa70d9394b6b99e8bd71b385ee3f057d2bf9344dd199"} Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.692398 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-f9b88fb79-vclnv" event={"ID":"77432545-f22c-453a-b6a7-7c932712efa9","Type":"ContainerStarted","Data":"3feb14a77d1251016f9a88a978c0277c39b5f555f9ba1d5f8c9fa84f835e2d6d"} Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.693849 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.693875 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.697135 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-a38c-account-create-update-x226p" event={"ID":"028d46f7-6f14-40f6-a6cf-77aeec59a99e","Type":"ContainerStarted","Data":"c62adef6e5f9bf0990f5fcc77903287bc27cc215bf5cbcd7af08a7afb5a9a0dc"} Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.698168 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-zfz4w" event={"ID":"87e82da2-fdd4-4ead-8a9b-022e55a03690","Type":"ContainerStarted","Data":"ecc4e3a464591c3ae12895248b97be713619753b60d765c32197f20eefb25c74"} Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.699479 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"bb5a0456-b5c5-433a-afde-fe38740e2310","Type":"ContainerStarted","Data":"a33772e7a61f00a233eb59cbe3495ba3e1ac72e11bed9784618da270828a271e"} Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.703345 4775 generic.go:334] "Generic (PLEG): container finished" podID="6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4" containerID="06c961a967b1c07bf826487fd3c050eb92ff3106f515b9136df5283020977b92" exitCode=0 Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.703584 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68ccf5bf68-lf5dz" event={"ID":"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4","Type":"ContainerDied","Data":"06c961a967b1c07bf826487fd3c050eb92ff3106f515b9136df5283020977b92"} Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.709254 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-tmm94" event={"ID":"bc238c51-34cc-4cd6-bdf2-00e3747a9d67","Type":"ContainerStarted","Data":"a900d64d00970943729e3dd5370a47455bba6366eba49077571ff7b009d713fa"} Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.711906 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-f9b88fb79-vclnv" podUID="77432545-f22c-453a-b6a7-7c932712efa9" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.713092 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db3c-account-create-update-k29zp" event={"ID":"22a5e145-5b67-4d02-9104-ade3e48888db","Type":"ContainerStarted","Data":"eb428452fe127829b5fdd3edea1e21f849f0fb951673f81da2334a3d27807ded"} Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.713136 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db3c-account-create-update-k29zp" event={"ID":"22a5e145-5b67-4d02-9104-ade3e48888db","Type":"ContainerStarted","Data":"e609ddd387c7c54bd1869b55edbc31d421deac134c26c25fe6a20a50afdcea02"} Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.744239 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-45fb-account-create-update-glct2" event={"ID":"330e8d60-7af3-42ed-a0d7-234264037d09","Type":"ContainerStarted","Data":"e350cae4a6e948bcd9c31418f7d52ae8e550001eb6a2f7a579e8a1278c01e74c"} Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.745409 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-f9b88fb79-vclnv" podStartSLOduration=12.745394242 podStartE2EDuration="12.745394242s" podCreationTimestamp="2026-03-21 05:08:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:08:26.717530435 +0000 UTC m=+1259.693994049" watchObservedRunningTime="2026-03-21 05:08:26.745394242 +0000 UTC m=+1259.721857866" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.750483 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db3c-account-create-update-k29zp" podStartSLOduration=9.750467826 podStartE2EDuration="9.750467826s" podCreationTimestamp="2026-03-21 05:08:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:08:26.747173453 +0000 UTC m=+1259.723637087" watchObservedRunningTime="2026-03-21 05:08:26.750467826 +0000 UTC m=+1259.726931450" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.796792 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.473224251 podStartE2EDuration="17.796769394s" podCreationTimestamp="2026-03-21 05:08:09 +0000 UTC" firstStartedPulling="2026-03-21 05:08:10.048841198 +0000 UTC m=+1243.025304822" lastFinishedPulling="2026-03-21 05:08:25.372386341 +0000 UTC m=+1258.348849965" observedRunningTime="2026-03-21 05:08:26.781081161 +0000 UTC m=+1259.757544775" watchObservedRunningTime="2026-03-21 05:08:26.796769394 +0000 UTC m=+1259.773233018" Mar 21 05:08:26 crc kubenswrapper[4775]: I0321 05:08:26.993039 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:08:27 crc kubenswrapper[4775]: I0321 05:08:27.011461 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 05:08:27 crc kubenswrapper[4775]: I0321 05:08:27.684800 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86abc8a1-1d5e-4dad-9e62-4f1c88608b95" path="/var/lib/kubelet/pods/86abc8a1-1d5e-4dad-9e62-4f1c88608b95/volumes" Mar 21 05:08:27 crc kubenswrapper[4775]: I0321 05:08:27.686586 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ac5c16c-56ac-4299-ae61-a8200986ce10" path="/var/lib/kubelet/pods/8ac5c16c-56ac-4299-ae61-a8200986ce10/volumes" Mar 21 05:08:27 crc kubenswrapper[4775]: I0321 05:08:27.755134 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-45fb-account-create-update-glct2" event={"ID":"330e8d60-7af3-42ed-a0d7-234264037d09","Type":"ContainerStarted","Data":"1fd0a588f9d34ae16dba50e92cfc4f254b48cad183d5ee54a522ba1c5c91886c"} Mar 21 05:08:27 crc kubenswrapper[4775]: I0321 05:08:27.757319 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3fcc332-1498-4e1d-bc39-58c203a4b6ad","Type":"ContainerStarted","Data":"03c1e47b1d8dd7264986151b452287f41c3c0c4a42e00d07ccce4db26c6d480e"} Mar 21 05:08:27 crc kubenswrapper[4775]: I0321 05:08:27.765895 4775 generic.go:334] "Generic (PLEG): container finished" podID="028d46f7-6f14-40f6-a6cf-77aeec59a99e" containerID="959f27ca1f8883c27a7d20345a9d58bad88607efdd52bc8596a92a7571fa819e" exitCode=0 Mar 21 05:08:27 crc kubenswrapper[4775]: I0321 05:08:27.765981 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-a38c-account-create-update-x226p" event={"ID":"028d46f7-6f14-40f6-a6cf-77aeec59a99e","Type":"ContainerDied","Data":"959f27ca1f8883c27a7d20345a9d58bad88607efdd52bc8596a92a7571fa819e"} Mar 21 05:08:27 crc kubenswrapper[4775]: I0321 05:08:27.770743 4775 generic.go:334] "Generic (PLEG): container finished" podID="87e82da2-fdd4-4ead-8a9b-022e55a03690" containerID="f761be7525081d117b389918983117196fdba3258fd2ecd1f48a78a479770d1c" exitCode=0 Mar 21 05:08:27 crc kubenswrapper[4775]: I0321 05:08:27.770797 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-zfz4w" event={"ID":"87e82da2-fdd4-4ead-8a9b-022e55a03690","Type":"ContainerDied","Data":"f761be7525081d117b389918983117196fdba3258fd2ecd1f48a78a479770d1c"} Mar 21 05:08:27 crc kubenswrapper[4775]: I0321 05:08:27.773633 4775 generic.go:334] "Generic (PLEG): container finished" podID="bc238c51-34cc-4cd6-bdf2-00e3747a9d67" containerID="647e5b30ad956a8a9a631808f860ad303cf32d4aa053654a5cdcf1129262af27" exitCode=0 Mar 21 05:08:27 crc kubenswrapper[4775]: I0321 05:08:27.773710 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-tmm94" event={"ID":"bc238c51-34cc-4cd6-bdf2-00e3747a9d67","Type":"ContainerDied","Data":"647e5b30ad956a8a9a631808f860ad303cf32d4aa053654a5cdcf1129262af27"} Mar 21 05:08:27 crc kubenswrapper[4775]: I0321 05:08:27.776517 4775 generic.go:334] "Generic (PLEG): container finished" podID="b65ed673-344a-4064-ae90-c9d20964c648" containerID="0ff1159b21db4525f027b0be154a99974ea3e42eac6bac07387f589a08d088ce" exitCode=0 Mar 21 05:08:27 crc kubenswrapper[4775]: I0321 05:08:27.776594 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wxmkg" event={"ID":"b65ed673-344a-4064-ae90-c9d20964c648","Type":"ContainerDied","Data":"0ff1159b21db4525f027b0be154a99974ea3e42eac6bac07387f589a08d088ce"} Mar 21 05:08:27 crc kubenswrapper[4775]: I0321 05:08:27.778318 4775 generic.go:334] "Generic (PLEG): container finished" podID="22a5e145-5b67-4d02-9104-ade3e48888db" containerID="eb428452fe127829b5fdd3edea1e21f849f0fb951673f81da2334a3d27807ded" exitCode=0 Mar 21 05:08:27 crc kubenswrapper[4775]: I0321 05:08:27.778410 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db3c-account-create-update-k29zp" event={"ID":"22a5e145-5b67-4d02-9104-ade3e48888db","Type":"ContainerDied","Data":"eb428452fe127829b5fdd3edea1e21f849f0fb951673f81da2334a3d27807ded"} Mar 21 05:08:27 crc kubenswrapper[4775]: I0321 05:08:27.793484 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:27 crc kubenswrapper[4775]: I0321 05:08:27.932690 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-45fb-account-create-update-glct2" podStartSLOduration=10.932671226 podStartE2EDuration="10.932671226s" podCreationTimestamp="2026-03-21 05:08:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:08:27.901530426 +0000 UTC m=+1260.877994060" watchObservedRunningTime="2026-03-21 05:08:27.932671226 +0000 UTC m=+1260.909134850" Mar 21 05:08:28 crc kubenswrapper[4775]: I0321 05:08:28.787989 4775 generic.go:334] "Generic (PLEG): container finished" podID="330e8d60-7af3-42ed-a0d7-234264037d09" containerID="1fd0a588f9d34ae16dba50e92cfc4f254b48cad183d5ee54a522ba1c5c91886c" exitCode=0 Mar 21 05:08:28 crc kubenswrapper[4775]: I0321 05:08:28.788065 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-45fb-account-create-update-glct2" event={"ID":"330e8d60-7af3-42ed-a0d7-234264037d09","Type":"ContainerDied","Data":"1fd0a588f9d34ae16dba50e92cfc4f254b48cad183d5ee54a522ba1c5c91886c"} Mar 21 05:08:28 crc kubenswrapper[4775]: I0321 05:08:28.790369 4775 generic.go:334] "Generic (PLEG): container finished" podID="6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4" containerID="f14a2e346ace37be895dcc94dd25c65e41716e7e754b07d2e1e3277674aeac90" exitCode=0 Mar 21 05:08:28 crc kubenswrapper[4775]: I0321 05:08:28.790487 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68ccf5bf68-lf5dz" event={"ID":"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4","Type":"ContainerDied","Data":"f14a2e346ace37be895dcc94dd25c65e41716e7e754b07d2e1e3277674aeac90"} Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.267747 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a38c-account-create-update-x226p" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.366573 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jcbnj\" (UniqueName: \"kubernetes.io/projected/028d46f7-6f14-40f6-a6cf-77aeec59a99e-kube-api-access-jcbnj\") pod \"028d46f7-6f14-40f6-a6cf-77aeec59a99e\" (UID: \"028d46f7-6f14-40f6-a6cf-77aeec59a99e\") " Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.367913 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/028d46f7-6f14-40f6-a6cf-77aeec59a99e-operator-scripts\") pod \"028d46f7-6f14-40f6-a6cf-77aeec59a99e\" (UID: \"028d46f7-6f14-40f6-a6cf-77aeec59a99e\") " Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.368429 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/028d46f7-6f14-40f6-a6cf-77aeec59a99e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "028d46f7-6f14-40f6-a6cf-77aeec59a99e" (UID: "028d46f7-6f14-40f6-a6cf-77aeec59a99e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.368920 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/028d46f7-6f14-40f6-a6cf-77aeec59a99e-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.370729 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/028d46f7-6f14-40f6-a6cf-77aeec59a99e-kube-api-access-jcbnj" (OuterVolumeSpecName: "kube-api-access-jcbnj") pod "028d46f7-6f14-40f6-a6cf-77aeec59a99e" (UID: "028d46f7-6f14-40f6-a6cf-77aeec59a99e"). InnerVolumeSpecName "kube-api-access-jcbnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.459936 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-zfz4w" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.472820 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db3c-account-create-update-k29zp" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.473593 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jcbnj\" (UniqueName: \"kubernetes.io/projected/028d46f7-6f14-40f6-a6cf-77aeec59a99e-kube-api-access-jcbnj\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.493916 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-tmm94" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.500334 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wxmkg" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.581710 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkh9x\" (UniqueName: \"kubernetes.io/projected/bc238c51-34cc-4cd6-bdf2-00e3747a9d67-kube-api-access-xkh9x\") pod \"bc238c51-34cc-4cd6-bdf2-00e3747a9d67\" (UID: \"bc238c51-34cc-4cd6-bdf2-00e3747a9d67\") " Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.581840 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txzb7\" (UniqueName: \"kubernetes.io/projected/87e82da2-fdd4-4ead-8a9b-022e55a03690-kube-api-access-txzb7\") pod \"87e82da2-fdd4-4ead-8a9b-022e55a03690\" (UID: \"87e82da2-fdd4-4ead-8a9b-022e55a03690\") " Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.581938 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22a5e145-5b67-4d02-9104-ade3e48888db-operator-scripts\") pod \"22a5e145-5b67-4d02-9104-ade3e48888db\" (UID: \"22a5e145-5b67-4d02-9104-ade3e48888db\") " Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.581971 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqfwp\" (UniqueName: \"kubernetes.io/projected/22a5e145-5b67-4d02-9104-ade3e48888db-kube-api-access-lqfwp\") pod \"22a5e145-5b67-4d02-9104-ade3e48888db\" (UID: \"22a5e145-5b67-4d02-9104-ade3e48888db\") " Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.582047 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc238c51-34cc-4cd6-bdf2-00e3747a9d67-operator-scripts\") pod \"bc238c51-34cc-4cd6-bdf2-00e3747a9d67\" (UID: \"bc238c51-34cc-4cd6-bdf2-00e3747a9d67\") " Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.582136 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87e82da2-fdd4-4ead-8a9b-022e55a03690-operator-scripts\") pod \"87e82da2-fdd4-4ead-8a9b-022e55a03690\" (UID: \"87e82da2-fdd4-4ead-8a9b-022e55a03690\") " Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.584213 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc238c51-34cc-4cd6-bdf2-00e3747a9d67-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bc238c51-34cc-4cd6-bdf2-00e3747a9d67" (UID: "bc238c51-34cc-4cd6-bdf2-00e3747a9d67"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.584353 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87e82da2-fdd4-4ead-8a9b-022e55a03690-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "87e82da2-fdd4-4ead-8a9b-022e55a03690" (UID: "87e82da2-fdd4-4ead-8a9b-022e55a03690"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.584818 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22a5e145-5b67-4d02-9104-ade3e48888db-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "22a5e145-5b67-4d02-9104-ade3e48888db" (UID: "22a5e145-5b67-4d02-9104-ade3e48888db"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.604939 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87e82da2-fdd4-4ead-8a9b-022e55a03690-kube-api-access-txzb7" (OuterVolumeSpecName: "kube-api-access-txzb7") pod "87e82da2-fdd4-4ead-8a9b-022e55a03690" (UID: "87e82da2-fdd4-4ead-8a9b-022e55a03690"). InnerVolumeSpecName "kube-api-access-txzb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.604988 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22a5e145-5b67-4d02-9104-ade3e48888db-kube-api-access-lqfwp" (OuterVolumeSpecName: "kube-api-access-lqfwp") pod "22a5e145-5b67-4d02-9104-ade3e48888db" (UID: "22a5e145-5b67-4d02-9104-ade3e48888db"). InnerVolumeSpecName "kube-api-access-lqfwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.611347 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc238c51-34cc-4cd6-bdf2-00e3747a9d67-kube-api-access-xkh9x" (OuterVolumeSpecName: "kube-api-access-xkh9x") pod "bc238c51-34cc-4cd6-bdf2-00e3747a9d67" (UID: "bc238c51-34cc-4cd6-bdf2-00e3747a9d67"). InnerVolumeSpecName "kube-api-access-xkh9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.683706 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b65ed673-344a-4064-ae90-c9d20964c648-operator-scripts\") pod \"b65ed673-344a-4064-ae90-c9d20964c648\" (UID: \"b65ed673-344a-4064-ae90-c9d20964c648\") " Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.684037 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb6wb\" (UniqueName: \"kubernetes.io/projected/b65ed673-344a-4064-ae90-c9d20964c648-kube-api-access-wb6wb\") pod \"b65ed673-344a-4064-ae90-c9d20964c648\" (UID: \"b65ed673-344a-4064-ae90-c9d20964c648\") " Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.684469 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87e82da2-fdd4-4ead-8a9b-022e55a03690-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.684485 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkh9x\" (UniqueName: \"kubernetes.io/projected/bc238c51-34cc-4cd6-bdf2-00e3747a9d67-kube-api-access-xkh9x\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.684494 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txzb7\" (UniqueName: \"kubernetes.io/projected/87e82da2-fdd4-4ead-8a9b-022e55a03690-kube-api-access-txzb7\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.684503 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22a5e145-5b67-4d02-9104-ade3e48888db-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.684512 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqfwp\" (UniqueName: \"kubernetes.io/projected/22a5e145-5b67-4d02-9104-ade3e48888db-kube-api-access-lqfwp\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.684520 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc238c51-34cc-4cd6-bdf2-00e3747a9d67-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.686767 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b65ed673-344a-4064-ae90-c9d20964c648-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b65ed673-344a-4064-ae90-c9d20964c648" (UID: "b65ed673-344a-4064-ae90-c9d20964c648"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.688413 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b65ed673-344a-4064-ae90-c9d20964c648-kube-api-access-wb6wb" (OuterVolumeSpecName: "kube-api-access-wb6wb") pod "b65ed673-344a-4064-ae90-c9d20964c648" (UID: "b65ed673-344a-4064-ae90-c9d20964c648"). InnerVolumeSpecName "kube-api-access-wb6wb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.689690 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.785594 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-combined-ca-bundle\") pod \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\" (UID: \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\") " Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.785687 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-ovndb-tls-certs\") pod \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\" (UID: \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\") " Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.785753 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-config\") pod \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\" (UID: \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\") " Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.785831 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-httpd-config\") pod \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\" (UID: \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\") " Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.785915 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv44z\" (UniqueName: \"kubernetes.io/projected/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-kube-api-access-bv44z\") pod \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\" (UID: \"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4\") " Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.786326 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb6wb\" (UniqueName: \"kubernetes.io/projected/b65ed673-344a-4064-ae90-c9d20964c648-kube-api-access-wb6wb\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.786344 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b65ed673-344a-4064-ae90-c9d20964c648-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.790355 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4" (UID: "6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.790941 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-kube-api-access-bv44z" (OuterVolumeSpecName: "kube-api-access-bv44z") pod "6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4" (UID: "6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4"). InnerVolumeSpecName "kube-api-access-bv44z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.809597 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-tmm94" event={"ID":"bc238c51-34cc-4cd6-bdf2-00e3747a9d67","Type":"ContainerDied","Data":"a900d64d00970943729e3dd5370a47455bba6366eba49077571ff7b009d713fa"} Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.809671 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a900d64d00970943729e3dd5370a47455bba6366eba49077571ff7b009d713fa" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.809759 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-tmm94" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.812500 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db3c-account-create-update-k29zp" event={"ID":"22a5e145-5b67-4d02-9104-ade3e48888db","Type":"ContainerDied","Data":"e609ddd387c7c54bd1869b55edbc31d421deac134c26c25fe6a20a50afdcea02"} Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.812550 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e609ddd387c7c54bd1869b55edbc31d421deac134c26c25fe6a20a50afdcea02" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.812592 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db3c-account-create-update-k29zp" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.814066 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3fcc332-1498-4e1d-bc39-58c203a4b6ad","Type":"ContainerStarted","Data":"15da433bbe56b39c4ebf0eff9011ebe47b1beb92e06cabf9cd3d182dd0437ac0"} Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.815867 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-a38c-account-create-update-x226p" event={"ID":"028d46f7-6f14-40f6-a6cf-77aeec59a99e","Type":"ContainerDied","Data":"c62adef6e5f9bf0990f5fcc77903287bc27cc215bf5cbcd7af08a7afb5a9a0dc"} Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.815904 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c62adef6e5f9bf0990f5fcc77903287bc27cc215bf5cbcd7af08a7afb5a9a0dc" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.815959 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a38c-account-create-update-x226p" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.818638 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-zfz4w" event={"ID":"87e82da2-fdd4-4ead-8a9b-022e55a03690","Type":"ContainerDied","Data":"ecc4e3a464591c3ae12895248b97be713619753b60d765c32197f20eefb25c74"} Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.818669 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ecc4e3a464591c3ae12895248b97be713619753b60d765c32197f20eefb25c74" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.818725 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-zfz4w" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.821455 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68ccf5bf68-lf5dz" event={"ID":"6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4","Type":"ContainerDied","Data":"785cf1f0611121bc184a4bc0f34377c346c490853673595b9763600cad5d4149"} Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.821498 4775 scope.go:117] "RemoveContainer" containerID="06c961a967b1c07bf826487fd3c050eb92ff3106f515b9136df5283020977b92" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.821626 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-68ccf5bf68-lf5dz" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.827663 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wxmkg" event={"ID":"b65ed673-344a-4064-ae90-c9d20964c648","Type":"ContainerDied","Data":"a48ddd114a87c3d3cbfffa70d9394b6b99e8bd71b385ee3f057d2bf9344dd199"} Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.827728 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a48ddd114a87c3d3cbfffa70d9394b6b99e8bd71b385ee3f057d2bf9344dd199" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.827814 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wxmkg" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.854287 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-config" (OuterVolumeSpecName: "config") pod "6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4" (UID: "6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.854430 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4" (UID: "6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.878549 4775 scope.go:117] "RemoveContainer" containerID="f14a2e346ace37be895dcc94dd25c65e41716e7e754b07d2e1e3277674aeac90" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.888543 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.888572 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.888589 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.888600 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv44z\" (UniqueName: \"kubernetes.io/projected/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-kube-api-access-bv44z\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.911828 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4" (UID: "6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:29 crc kubenswrapper[4775]: I0321 05:08:29.990733 4775 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:30 crc kubenswrapper[4775]: I0321 05:08:30.137103 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-45fb-account-create-update-glct2" Mar 21 05:08:30 crc kubenswrapper[4775]: I0321 05:08:30.243602 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-68ccf5bf68-lf5dz"] Mar 21 05:08:30 crc kubenswrapper[4775]: I0321 05:08:30.253877 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-68ccf5bf68-lf5dz"] Mar 21 05:08:30 crc kubenswrapper[4775]: I0321 05:08:30.299220 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/330e8d60-7af3-42ed-a0d7-234264037d09-operator-scripts\") pod \"330e8d60-7af3-42ed-a0d7-234264037d09\" (UID: \"330e8d60-7af3-42ed-a0d7-234264037d09\") " Mar 21 05:08:30 crc kubenswrapper[4775]: I0321 05:08:30.299614 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47c7f\" (UniqueName: \"kubernetes.io/projected/330e8d60-7af3-42ed-a0d7-234264037d09-kube-api-access-47c7f\") pod \"330e8d60-7af3-42ed-a0d7-234264037d09\" (UID: \"330e8d60-7af3-42ed-a0d7-234264037d09\") " Mar 21 05:08:30 crc kubenswrapper[4775]: I0321 05:08:30.299779 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/330e8d60-7af3-42ed-a0d7-234264037d09-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "330e8d60-7af3-42ed-a0d7-234264037d09" (UID: "330e8d60-7af3-42ed-a0d7-234264037d09"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:08:30 crc kubenswrapper[4775]: I0321 05:08:30.300395 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/330e8d60-7af3-42ed-a0d7-234264037d09-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:30 crc kubenswrapper[4775]: I0321 05:08:30.311898 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-f9b88fb79-vclnv" Mar 21 05:08:30 crc kubenswrapper[4775]: I0321 05:08:30.311981 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/330e8d60-7af3-42ed-a0d7-234264037d09-kube-api-access-47c7f" (OuterVolumeSpecName: "kube-api-access-47c7f") pod "330e8d60-7af3-42ed-a0d7-234264037d09" (UID: "330e8d60-7af3-42ed-a0d7-234264037d09"). InnerVolumeSpecName "kube-api-access-47c7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:08:30 crc kubenswrapper[4775]: I0321 05:08:30.401987 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47c7f\" (UniqueName: \"kubernetes.io/projected/330e8d60-7af3-42ed-a0d7-234264037d09-kube-api-access-47c7f\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:30 crc kubenswrapper[4775]: I0321 05:08:30.852020 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3fcc332-1498-4e1d-bc39-58c203a4b6ad","Type":"ContainerStarted","Data":"59ff223fce72d0f0a5dcf748449e4af7793a03edd914b1484e21b7aca4e40d65"} Mar 21 05:08:30 crc kubenswrapper[4775]: I0321 05:08:30.880594 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-45fb-account-create-update-glct2" Mar 21 05:08:30 crc kubenswrapper[4775]: I0321 05:08:30.880870 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-45fb-account-create-update-glct2" event={"ID":"330e8d60-7af3-42ed-a0d7-234264037d09","Type":"ContainerDied","Data":"e350cae4a6e948bcd9c31418f7d52ae8e550001eb6a2f7a579e8a1278c01e74c"} Mar 21 05:08:30 crc kubenswrapper[4775]: I0321 05:08:30.881316 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e350cae4a6e948bcd9c31418f7d52ae8e550001eb6a2f7a579e8a1278c01e74c" Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.335545 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.425189 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/09cf3763-0b41-4452-a247-d9a56f58b05d-horizon-tls-certs\") pod \"09cf3763-0b41-4452-a247-d9a56f58b05d\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.425537 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09cf3763-0b41-4452-a247-d9a56f58b05d-config-data\") pod \"09cf3763-0b41-4452-a247-d9a56f58b05d\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.425590 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09cf3763-0b41-4452-a247-d9a56f58b05d-combined-ca-bundle\") pod \"09cf3763-0b41-4452-a247-d9a56f58b05d\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.425695 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09cf3763-0b41-4452-a247-d9a56f58b05d-logs\") pod \"09cf3763-0b41-4452-a247-d9a56f58b05d\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.425713 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rg4gg\" (UniqueName: \"kubernetes.io/projected/09cf3763-0b41-4452-a247-d9a56f58b05d-kube-api-access-rg4gg\") pod \"09cf3763-0b41-4452-a247-d9a56f58b05d\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.425745 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09cf3763-0b41-4452-a247-d9a56f58b05d-scripts\") pod \"09cf3763-0b41-4452-a247-d9a56f58b05d\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.426251 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09cf3763-0b41-4452-a247-d9a56f58b05d-logs" (OuterVolumeSpecName: "logs") pod "09cf3763-0b41-4452-a247-d9a56f58b05d" (UID: "09cf3763-0b41-4452-a247-d9a56f58b05d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.426407 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/09cf3763-0b41-4452-a247-d9a56f58b05d-horizon-secret-key\") pod \"09cf3763-0b41-4452-a247-d9a56f58b05d\" (UID: \"09cf3763-0b41-4452-a247-d9a56f58b05d\") " Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.426834 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09cf3763-0b41-4452-a247-d9a56f58b05d-logs\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.430208 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09cf3763-0b41-4452-a247-d9a56f58b05d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "09cf3763-0b41-4452-a247-d9a56f58b05d" (UID: "09cf3763-0b41-4452-a247-d9a56f58b05d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.434939 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09cf3763-0b41-4452-a247-d9a56f58b05d-kube-api-access-rg4gg" (OuterVolumeSpecName: "kube-api-access-rg4gg") pod "09cf3763-0b41-4452-a247-d9a56f58b05d" (UID: "09cf3763-0b41-4452-a247-d9a56f58b05d"). InnerVolumeSpecName "kube-api-access-rg4gg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.456173 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09cf3763-0b41-4452-a247-d9a56f58b05d-scripts" (OuterVolumeSpecName: "scripts") pod "09cf3763-0b41-4452-a247-d9a56f58b05d" (UID: "09cf3763-0b41-4452-a247-d9a56f58b05d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.458027 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09cf3763-0b41-4452-a247-d9a56f58b05d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "09cf3763-0b41-4452-a247-d9a56f58b05d" (UID: "09cf3763-0b41-4452-a247-d9a56f58b05d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.459870 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09cf3763-0b41-4452-a247-d9a56f58b05d-config-data" (OuterVolumeSpecName: "config-data") pod "09cf3763-0b41-4452-a247-d9a56f58b05d" (UID: "09cf3763-0b41-4452-a247-d9a56f58b05d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.485628 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09cf3763-0b41-4452-a247-d9a56f58b05d-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "09cf3763-0b41-4452-a247-d9a56f58b05d" (UID: "09cf3763-0b41-4452-a247-d9a56f58b05d"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.528887 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rg4gg\" (UniqueName: \"kubernetes.io/projected/09cf3763-0b41-4452-a247-d9a56f58b05d-kube-api-access-rg4gg\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.528930 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09cf3763-0b41-4452-a247-d9a56f58b05d-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.528941 4775 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/09cf3763-0b41-4452-a247-d9a56f58b05d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.528949 4775 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/09cf3763-0b41-4452-a247-d9a56f58b05d-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.528957 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09cf3763-0b41-4452-a247-d9a56f58b05d-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.528965 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09cf3763-0b41-4452-a247-d9a56f58b05d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.672837 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4" path="/var/lib/kubelet/pods/6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4/volumes" Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.892063 4775 generic.go:334] "Generic (PLEG): container finished" podID="09cf3763-0b41-4452-a247-d9a56f58b05d" containerID="4df1223c20d1958e562bffafa09f6a9587b632ec38a4f1dbf6a26792e8ce21fd" exitCode=137 Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.892141 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64fb567758-hd2ld" Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.892971 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64fb567758-hd2ld" event={"ID":"09cf3763-0b41-4452-a247-d9a56f58b05d","Type":"ContainerDied","Data":"4df1223c20d1958e562bffafa09f6a9587b632ec38a4f1dbf6a26792e8ce21fd"} Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.893062 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64fb567758-hd2ld" event={"ID":"09cf3763-0b41-4452-a247-d9a56f58b05d","Type":"ContainerDied","Data":"7045bd948f55810ab12433655a5f2808bafab5e3e89d9d68cb00186c737c1405"} Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.893095 4775 scope.go:117] "RemoveContainer" containerID="1da6670625f81794dc89cffb06ac6d008a9c7c8ed1b16d49c09b2332ff0e94fa" Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.894438 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3fcc332-1498-4e1d-bc39-58c203a4b6ad","Type":"ContainerStarted","Data":"6118bee8bc5688dea0b47263f9e7450128104ab9d8f36622df267e16e2ba897f"} Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.915937 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-64fb567758-hd2ld"] Mar 21 05:08:31 crc kubenswrapper[4775]: I0321 05:08:31.931086 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-64fb567758-hd2ld"] Mar 21 05:08:32 crc kubenswrapper[4775]: I0321 05:08:32.059407 4775 scope.go:117] "RemoveContainer" containerID="4df1223c20d1958e562bffafa09f6a9587b632ec38a4f1dbf6a26792e8ce21fd" Mar 21 05:08:32 crc kubenswrapper[4775]: I0321 05:08:32.084702 4775 scope.go:117] "RemoveContainer" containerID="1da6670625f81794dc89cffb06ac6d008a9c7c8ed1b16d49c09b2332ff0e94fa" Mar 21 05:08:32 crc kubenswrapper[4775]: E0321 05:08:32.085257 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1da6670625f81794dc89cffb06ac6d008a9c7c8ed1b16d49c09b2332ff0e94fa\": container with ID starting with 1da6670625f81794dc89cffb06ac6d008a9c7c8ed1b16d49c09b2332ff0e94fa not found: ID does not exist" containerID="1da6670625f81794dc89cffb06ac6d008a9c7c8ed1b16d49c09b2332ff0e94fa" Mar 21 05:08:32 crc kubenswrapper[4775]: I0321 05:08:32.085293 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1da6670625f81794dc89cffb06ac6d008a9c7c8ed1b16d49c09b2332ff0e94fa"} err="failed to get container status \"1da6670625f81794dc89cffb06ac6d008a9c7c8ed1b16d49c09b2332ff0e94fa\": rpc error: code = NotFound desc = could not find container \"1da6670625f81794dc89cffb06ac6d008a9c7c8ed1b16d49c09b2332ff0e94fa\": container with ID starting with 1da6670625f81794dc89cffb06ac6d008a9c7c8ed1b16d49c09b2332ff0e94fa not found: ID does not exist" Mar 21 05:08:32 crc kubenswrapper[4775]: I0321 05:08:32.085319 4775 scope.go:117] "RemoveContainer" containerID="4df1223c20d1958e562bffafa09f6a9587b632ec38a4f1dbf6a26792e8ce21fd" Mar 21 05:08:32 crc kubenswrapper[4775]: E0321 05:08:32.085549 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4df1223c20d1958e562bffafa09f6a9587b632ec38a4f1dbf6a26792e8ce21fd\": container with ID starting with 4df1223c20d1958e562bffafa09f6a9587b632ec38a4f1dbf6a26792e8ce21fd not found: ID does not exist" containerID="4df1223c20d1958e562bffafa09f6a9587b632ec38a4f1dbf6a26792e8ce21fd" Mar 21 05:08:32 crc kubenswrapper[4775]: I0321 05:08:32.085644 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4df1223c20d1958e562bffafa09f6a9587b632ec38a4f1dbf6a26792e8ce21fd"} err="failed to get container status \"4df1223c20d1958e562bffafa09f6a9587b632ec38a4f1dbf6a26792e8ce21fd\": rpc error: code = NotFound desc = could not find container \"4df1223c20d1958e562bffafa09f6a9587b632ec38a4f1dbf6a26792e8ce21fd\": container with ID starting with 4df1223c20d1958e562bffafa09f6a9587b632ec38a4f1dbf6a26792e8ce21fd not found: ID does not exist" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.179310 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gk78h"] Mar 21 05:08:33 crc kubenswrapper[4775]: E0321 05:08:33.180453 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="330e8d60-7af3-42ed-a0d7-234264037d09" containerName="mariadb-account-create-update" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.180466 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="330e8d60-7af3-42ed-a0d7-234264037d09" containerName="mariadb-account-create-update" Mar 21 05:08:33 crc kubenswrapper[4775]: E0321 05:08:33.180497 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22a5e145-5b67-4d02-9104-ade3e48888db" containerName="mariadb-account-create-update" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.180504 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="22a5e145-5b67-4d02-9104-ade3e48888db" containerName="mariadb-account-create-update" Mar 21 05:08:33 crc kubenswrapper[4775]: E0321 05:08:33.180518 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4" containerName="neutron-httpd" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.180525 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4" containerName="neutron-httpd" Mar 21 05:08:33 crc kubenswrapper[4775]: E0321 05:08:33.180537 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09cf3763-0b41-4452-a247-d9a56f58b05d" containerName="horizon" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.180544 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="09cf3763-0b41-4452-a247-d9a56f58b05d" containerName="horizon" Mar 21 05:08:33 crc kubenswrapper[4775]: E0321 05:08:33.180553 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="028d46f7-6f14-40f6-a6cf-77aeec59a99e" containerName="mariadb-account-create-update" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.180562 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="028d46f7-6f14-40f6-a6cf-77aeec59a99e" containerName="mariadb-account-create-update" Mar 21 05:08:33 crc kubenswrapper[4775]: E0321 05:08:33.180578 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87e82da2-fdd4-4ead-8a9b-022e55a03690" containerName="mariadb-database-create" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.180584 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="87e82da2-fdd4-4ead-8a9b-022e55a03690" containerName="mariadb-database-create" Mar 21 05:08:33 crc kubenswrapper[4775]: E0321 05:08:33.180606 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4" containerName="neutron-api" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.180613 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4" containerName="neutron-api" Mar 21 05:08:33 crc kubenswrapper[4775]: E0321 05:08:33.180635 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09cf3763-0b41-4452-a247-d9a56f58b05d" containerName="horizon-log" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.180641 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="09cf3763-0b41-4452-a247-d9a56f58b05d" containerName="horizon-log" Mar 21 05:08:33 crc kubenswrapper[4775]: E0321 05:08:33.180655 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc238c51-34cc-4cd6-bdf2-00e3747a9d67" containerName="mariadb-database-create" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.180662 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc238c51-34cc-4cd6-bdf2-00e3747a9d67" containerName="mariadb-database-create" Mar 21 05:08:33 crc kubenswrapper[4775]: E0321 05:08:33.180690 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b65ed673-344a-4064-ae90-c9d20964c648" containerName="mariadb-database-create" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.180696 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b65ed673-344a-4064-ae90-c9d20964c648" containerName="mariadb-database-create" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.180977 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b65ed673-344a-4064-ae90-c9d20964c648" containerName="mariadb-database-create" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.180995 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc238c51-34cc-4cd6-bdf2-00e3747a9d67" containerName="mariadb-database-create" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.181012 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="09cf3763-0b41-4452-a247-d9a56f58b05d" containerName="horizon" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.181027 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="09cf3763-0b41-4452-a247-d9a56f58b05d" containerName="horizon-log" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.181049 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="87e82da2-fdd4-4ead-8a9b-022e55a03690" containerName="mariadb-database-create" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.181074 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="330e8d60-7af3-42ed-a0d7-234264037d09" containerName="mariadb-account-create-update" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.181093 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="028d46f7-6f14-40f6-a6cf-77aeec59a99e" containerName="mariadb-account-create-update" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.181136 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4" containerName="neutron-httpd" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.181159 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dc5e1ad-635a-478c-8fac-ac1fcdb2bad4" containerName="neutron-api" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.181179 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="22a5e145-5b67-4d02-9104-ade3e48888db" containerName="mariadb-account-create-update" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.181890 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gk78h" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.187784 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bhr6z" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.188015 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.190172 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.206791 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gk78h"] Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.265053 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/824625b1-30cc-42c5-ad83-5854770c2f46-scripts\") pod \"nova-cell0-conductor-db-sync-gk78h\" (UID: \"824625b1-30cc-42c5-ad83-5854770c2f46\") " pod="openstack/nova-cell0-conductor-db-sync-gk78h" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.265204 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/824625b1-30cc-42c5-ad83-5854770c2f46-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-gk78h\" (UID: \"824625b1-30cc-42c5-ad83-5854770c2f46\") " pod="openstack/nova-cell0-conductor-db-sync-gk78h" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.265237 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59jrc\" (UniqueName: \"kubernetes.io/projected/824625b1-30cc-42c5-ad83-5854770c2f46-kube-api-access-59jrc\") pod \"nova-cell0-conductor-db-sync-gk78h\" (UID: \"824625b1-30cc-42c5-ad83-5854770c2f46\") " pod="openstack/nova-cell0-conductor-db-sync-gk78h" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.265285 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/824625b1-30cc-42c5-ad83-5854770c2f46-config-data\") pod \"nova-cell0-conductor-db-sync-gk78h\" (UID: \"824625b1-30cc-42c5-ad83-5854770c2f46\") " pod="openstack/nova-cell0-conductor-db-sync-gk78h" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.366777 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/824625b1-30cc-42c5-ad83-5854770c2f46-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-gk78h\" (UID: \"824625b1-30cc-42c5-ad83-5854770c2f46\") " pod="openstack/nova-cell0-conductor-db-sync-gk78h" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.366869 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59jrc\" (UniqueName: \"kubernetes.io/projected/824625b1-30cc-42c5-ad83-5854770c2f46-kube-api-access-59jrc\") pod \"nova-cell0-conductor-db-sync-gk78h\" (UID: \"824625b1-30cc-42c5-ad83-5854770c2f46\") " pod="openstack/nova-cell0-conductor-db-sync-gk78h" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.366928 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/824625b1-30cc-42c5-ad83-5854770c2f46-config-data\") pod \"nova-cell0-conductor-db-sync-gk78h\" (UID: \"824625b1-30cc-42c5-ad83-5854770c2f46\") " pod="openstack/nova-cell0-conductor-db-sync-gk78h" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.367024 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/824625b1-30cc-42c5-ad83-5854770c2f46-scripts\") pod \"nova-cell0-conductor-db-sync-gk78h\" (UID: \"824625b1-30cc-42c5-ad83-5854770c2f46\") " pod="openstack/nova-cell0-conductor-db-sync-gk78h" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.371887 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/824625b1-30cc-42c5-ad83-5854770c2f46-scripts\") pod \"nova-cell0-conductor-db-sync-gk78h\" (UID: \"824625b1-30cc-42c5-ad83-5854770c2f46\") " pod="openstack/nova-cell0-conductor-db-sync-gk78h" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.374649 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/824625b1-30cc-42c5-ad83-5854770c2f46-config-data\") pod \"nova-cell0-conductor-db-sync-gk78h\" (UID: \"824625b1-30cc-42c5-ad83-5854770c2f46\") " pod="openstack/nova-cell0-conductor-db-sync-gk78h" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.375005 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/824625b1-30cc-42c5-ad83-5854770c2f46-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-gk78h\" (UID: \"824625b1-30cc-42c5-ad83-5854770c2f46\") " pod="openstack/nova-cell0-conductor-db-sync-gk78h" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.392241 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59jrc\" (UniqueName: \"kubernetes.io/projected/824625b1-30cc-42c5-ad83-5854770c2f46-kube-api-access-59jrc\") pod \"nova-cell0-conductor-db-sync-gk78h\" (UID: \"824625b1-30cc-42c5-ad83-5854770c2f46\") " pod="openstack/nova-cell0-conductor-db-sync-gk78h" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.506875 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gk78h" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.560240 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.687794 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09cf3763-0b41-4452-a247-d9a56f58b05d" path="/var/lib/kubelet/pods/09cf3763-0b41-4452-a247-d9a56f58b05d/volumes" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.943157 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gk78h"] Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.957344 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3fcc332-1498-4e1d-bc39-58c203a4b6ad","Type":"ContainerStarted","Data":"07f243e539099e70b23a12537cd33bb12a02581f84122079a9f29395c247bf3f"} Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.958440 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 21 05:08:33 crc kubenswrapper[4775]: I0321 05:08:33.989667 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.087781939 podStartE2EDuration="7.989647859s" podCreationTimestamp="2026-03-21 05:08:26 +0000 UTC" firstStartedPulling="2026-03-21 05:08:27.011243565 +0000 UTC m=+1259.987707189" lastFinishedPulling="2026-03-21 05:08:32.913109495 +0000 UTC m=+1265.889573109" observedRunningTime="2026-03-21 05:08:33.98226302 +0000 UTC m=+1266.958726644" watchObservedRunningTime="2026-03-21 05:08:33.989647859 +0000 UTC m=+1266.966111483" Mar 21 05:08:34 crc kubenswrapper[4775]: I0321 05:08:34.969356 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gk78h" event={"ID":"824625b1-30cc-42c5-ad83-5854770c2f46","Type":"ContainerStarted","Data":"2d77ca12f9d9586582c849b2cbdaffac8b8dcfcc0b7a10cdf251cde2b34e8f90"} Mar 21 05:08:34 crc kubenswrapper[4775]: I0321 05:08:34.970063 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" containerName="ceilometer-central-agent" containerID="cri-o://15da433bbe56b39c4ebf0eff9011ebe47b1beb92e06cabf9cd3d182dd0437ac0" gracePeriod=30 Mar 21 05:08:34 crc kubenswrapper[4775]: I0321 05:08:34.970895 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" containerName="proxy-httpd" containerID="cri-o://07f243e539099e70b23a12537cd33bb12a02581f84122079a9f29395c247bf3f" gracePeriod=30 Mar 21 05:08:34 crc kubenswrapper[4775]: I0321 05:08:34.970889 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" containerName="sg-core" containerID="cri-o://6118bee8bc5688dea0b47263f9e7450128104ab9d8f36622df267e16e2ba897f" gracePeriod=30 Mar 21 05:08:34 crc kubenswrapper[4775]: I0321 05:08:34.970924 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" containerName="ceilometer-notification-agent" containerID="cri-o://59ff223fce72d0f0a5dcf748449e4af7793a03edd914b1484e21b7aca4e40d65" gracePeriod=30 Mar 21 05:08:35 crc kubenswrapper[4775]: I0321 05:08:35.982750 4775 generic.go:334] "Generic (PLEG): container finished" podID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" containerID="07f243e539099e70b23a12537cd33bb12a02581f84122079a9f29395c247bf3f" exitCode=0 Mar 21 05:08:35 crc kubenswrapper[4775]: I0321 05:08:35.983100 4775 generic.go:334] "Generic (PLEG): container finished" podID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" containerID="6118bee8bc5688dea0b47263f9e7450128104ab9d8f36622df267e16e2ba897f" exitCode=2 Mar 21 05:08:35 crc kubenswrapper[4775]: I0321 05:08:35.983148 4775 generic.go:334] "Generic (PLEG): container finished" podID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" containerID="59ff223fce72d0f0a5dcf748449e4af7793a03edd914b1484e21b7aca4e40d65" exitCode=0 Mar 21 05:08:35 crc kubenswrapper[4775]: I0321 05:08:35.982832 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3fcc332-1498-4e1d-bc39-58c203a4b6ad","Type":"ContainerDied","Data":"07f243e539099e70b23a12537cd33bb12a02581f84122079a9f29395c247bf3f"} Mar 21 05:08:35 crc kubenswrapper[4775]: I0321 05:08:35.983370 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3fcc332-1498-4e1d-bc39-58c203a4b6ad","Type":"ContainerDied","Data":"6118bee8bc5688dea0b47263f9e7450128104ab9d8f36622df267e16e2ba897f"} Mar 21 05:08:35 crc kubenswrapper[4775]: I0321 05:08:35.983388 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3fcc332-1498-4e1d-bc39-58c203a4b6ad","Type":"ContainerDied","Data":"59ff223fce72d0f0a5dcf748449e4af7793a03edd914b1484e21b7aca4e40d65"} Mar 21 05:08:37 crc kubenswrapper[4775]: I0321 05:08:37.949363 4775 scope.go:117] "RemoveContainer" containerID="a9b0f66cd072ab055301f7bd9fc0725dbbd203bf7fa9a48d07d610fb2904d30b" Mar 21 05:08:38 crc kubenswrapper[4775]: I0321 05:08:38.004521 4775 generic.go:334] "Generic (PLEG): container finished" podID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" containerID="15da433bbe56b39c4ebf0eff9011ebe47b1beb92e06cabf9cd3d182dd0437ac0" exitCode=0 Mar 21 05:08:38 crc kubenswrapper[4775]: I0321 05:08:38.004562 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3fcc332-1498-4e1d-bc39-58c203a4b6ad","Type":"ContainerDied","Data":"15da433bbe56b39c4ebf0eff9011ebe47b1beb92e06cabf9cd3d182dd0437ac0"} Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.174848 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.253320 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-sg-core-conf-yaml\") pod \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.253377 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-run-httpd\") pod \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.253780 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d3fcc332-1498-4e1d-bc39-58c203a4b6ad" (UID: "d3fcc332-1498-4e1d-bc39-58c203a4b6ad"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.282072 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d3fcc332-1498-4e1d-bc39-58c203a4b6ad" (UID: "d3fcc332-1498-4e1d-bc39-58c203a4b6ad"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.354885 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cmhz\" (UniqueName: \"kubernetes.io/projected/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-kube-api-access-8cmhz\") pod \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.354928 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-config-data\") pod \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.354995 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-combined-ca-bundle\") pod \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.355038 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-scripts\") pod \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.355075 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-log-httpd\") pod \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\" (UID: \"d3fcc332-1498-4e1d-bc39-58c203a4b6ad\") " Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.355463 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.355488 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.355954 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d3fcc332-1498-4e1d-bc39-58c203a4b6ad" (UID: "d3fcc332-1498-4e1d-bc39-58c203a4b6ad"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.367399 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-kube-api-access-8cmhz" (OuterVolumeSpecName: "kube-api-access-8cmhz") pod "d3fcc332-1498-4e1d-bc39-58c203a4b6ad" (UID: "d3fcc332-1498-4e1d-bc39-58c203a4b6ad"). InnerVolumeSpecName "kube-api-access-8cmhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.370751 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-scripts" (OuterVolumeSpecName: "scripts") pod "d3fcc332-1498-4e1d-bc39-58c203a4b6ad" (UID: "d3fcc332-1498-4e1d-bc39-58c203a4b6ad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.457612 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cmhz\" (UniqueName: \"kubernetes.io/projected/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-kube-api-access-8cmhz\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.457959 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.457977 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.468189 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3fcc332-1498-4e1d-bc39-58c203a4b6ad" (UID: "d3fcc332-1498-4e1d-bc39-58c203a4b6ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.480678 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-config-data" (OuterVolumeSpecName: "config-data") pod "d3fcc332-1498-4e1d-bc39-58c203a4b6ad" (UID: "d3fcc332-1498-4e1d-bc39-58c203a4b6ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.558901 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:42 crc kubenswrapper[4775]: I0321 05:08:42.558941 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3fcc332-1498-4e1d-bc39-58c203a4b6ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.081865 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gk78h" event={"ID":"824625b1-30cc-42c5-ad83-5854770c2f46","Type":"ContainerStarted","Data":"20d707fa7f70310d9cb80b150c04d331c2047a421ec95e9ba298af9cad00d19d"} Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.085560 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3fcc332-1498-4e1d-bc39-58c203a4b6ad","Type":"ContainerDied","Data":"03c1e47b1d8dd7264986151b452287f41c3c0c4a42e00d07ccce4db26c6d480e"} Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.085674 4775 scope.go:117] "RemoveContainer" containerID="07f243e539099e70b23a12537cd33bb12a02581f84122079a9f29395c247bf3f" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.085843 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.114605 4775 scope.go:117] "RemoveContainer" containerID="6118bee8bc5688dea0b47263f9e7450128104ab9d8f36622df267e16e2ba897f" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.132917 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-gk78h" podStartSLOduration=1.900723777 podStartE2EDuration="10.132897913s" podCreationTimestamp="2026-03-21 05:08:33 +0000 UTC" firstStartedPulling="2026-03-21 05:08:33.946712896 +0000 UTC m=+1266.923176520" lastFinishedPulling="2026-03-21 05:08:42.178887032 +0000 UTC m=+1275.155350656" observedRunningTime="2026-03-21 05:08:43.099986923 +0000 UTC m=+1276.076450587" watchObservedRunningTime="2026-03-21 05:08:43.132897913 +0000 UTC m=+1276.109361537" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.135618 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.142318 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.154614 4775 scope.go:117] "RemoveContainer" containerID="59ff223fce72d0f0a5dcf748449e4af7793a03edd914b1484e21b7aca4e40d65" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.156233 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:08:43 crc kubenswrapper[4775]: E0321 05:08:43.156753 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" containerName="proxy-httpd" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.156774 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" containerName="proxy-httpd" Mar 21 05:08:43 crc kubenswrapper[4775]: E0321 05:08:43.156788 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" containerName="ceilometer-central-agent" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.156794 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" containerName="ceilometer-central-agent" Mar 21 05:08:43 crc kubenswrapper[4775]: E0321 05:08:43.156841 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" containerName="sg-core" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.156847 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" containerName="sg-core" Mar 21 05:08:43 crc kubenswrapper[4775]: E0321 05:08:43.156859 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" containerName="ceilometer-notification-agent" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.156864 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" containerName="ceilometer-notification-agent" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.157059 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" containerName="ceilometer-notification-agent" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.157080 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" containerName="ceilometer-central-agent" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.157090 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" containerName="proxy-httpd" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.157100 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" containerName="sg-core" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.158797 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.164466 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.164771 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.172204 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.172273 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab91b8d2-32c9-4542-a730-7fc84e9892bd-log-httpd\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.172307 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.172361 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab91b8d2-32c9-4542-a730-7fc84e9892bd-run-httpd\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.172386 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4msqv\" (UniqueName: \"kubernetes.io/projected/ab91b8d2-32c9-4542-a730-7fc84e9892bd-kube-api-access-4msqv\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.172430 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-config-data\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.172449 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-scripts\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.172498 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.209380 4775 scope.go:117] "RemoveContainer" containerID="15da433bbe56b39c4ebf0eff9011ebe47b1beb92e06cabf9cd3d182dd0437ac0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.273974 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab91b8d2-32c9-4542-a730-7fc84e9892bd-run-httpd\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.274020 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4msqv\" (UniqueName: \"kubernetes.io/projected/ab91b8d2-32c9-4542-a730-7fc84e9892bd-kube-api-access-4msqv\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.274417 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab91b8d2-32c9-4542-a730-7fc84e9892bd-run-httpd\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.274515 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-config-data\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.274633 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-scripts\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.274801 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.275039 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab91b8d2-32c9-4542-a730-7fc84e9892bd-log-httpd\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.275486 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.275434 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab91b8d2-32c9-4542-a730-7fc84e9892bd-log-httpd\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.278518 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.279735 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-scripts\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.279952 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.280527 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-config-data\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.298058 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4msqv\" (UniqueName: \"kubernetes.io/projected/ab91b8d2-32c9-4542-a730-7fc84e9892bd-kube-api-access-4msqv\") pod \"ceilometer-0\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.478187 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.681742 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3fcc332-1498-4e1d-bc39-58c203a4b6ad" path="/var/lib/kubelet/pods/d3fcc332-1498-4e1d-bc39-58c203a4b6ad/volumes" Mar 21 05:08:43 crc kubenswrapper[4775]: I0321 05:08:43.950957 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:08:44 crc kubenswrapper[4775]: I0321 05:08:44.095541 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab91b8d2-32c9-4542-a730-7fc84e9892bd","Type":"ContainerStarted","Data":"086d1b17c5dfcb625f7ec04d8ff48e78b1e6cd139ff3016cea3aa265e3ae9e38"} Mar 21 05:08:44 crc kubenswrapper[4775]: I0321 05:08:44.745278 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 05:08:44 crc kubenswrapper[4775]: I0321 05:08:44.745744 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="47e577d7-e389-4135-b4fb-979bd627eaa9" containerName="glance-log" containerID="cri-o://29a18bd4308c28984ec8d5e5e490324db58708c1d07372b24ee7783e1995157d" gracePeriod=30 Mar 21 05:08:44 crc kubenswrapper[4775]: I0321 05:08:44.745818 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="47e577d7-e389-4135-b4fb-979bd627eaa9" containerName="glance-httpd" containerID="cri-o://0173c3dcb2f1ee362b645abaeead4a322f137b6409341a3b746ad5f19cffafb5" gracePeriod=30 Mar 21 05:08:45 crc kubenswrapper[4775]: I0321 05:08:45.108995 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab91b8d2-32c9-4542-a730-7fc84e9892bd","Type":"ContainerStarted","Data":"f3b6a0d856af07fa3386647a1df14c43107fa18d16407653836fe8f6805030db"} Mar 21 05:08:45 crc kubenswrapper[4775]: I0321 05:08:45.111162 4775 generic.go:334] "Generic (PLEG): container finished" podID="47e577d7-e389-4135-b4fb-979bd627eaa9" containerID="29a18bd4308c28984ec8d5e5e490324db58708c1d07372b24ee7783e1995157d" exitCode=143 Mar 21 05:08:45 crc kubenswrapper[4775]: I0321 05:08:45.111206 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47e577d7-e389-4135-b4fb-979bd627eaa9","Type":"ContainerDied","Data":"29a18bd4308c28984ec8d5e5e490324db58708c1d07372b24ee7783e1995157d"} Mar 21 05:08:46 crc kubenswrapper[4775]: I0321 05:08:46.121495 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab91b8d2-32c9-4542-a730-7fc84e9892bd","Type":"ContainerStarted","Data":"8cc8d28938a70addf2d6412cf322bdf56956d1adec39274fbeadbaacd685266d"} Mar 21 05:08:46 crc kubenswrapper[4775]: I0321 05:08:46.319343 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 05:08:46 crc kubenswrapper[4775]: I0321 05:08:46.319854 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="71255c4f-1e47-4e35-845f-876fff5fd6d4" containerName="glance-log" containerID="cri-o://b0bf1b828206099564568160f6163284ee71379761dd593fcfafe3139a339268" gracePeriod=30 Mar 21 05:08:46 crc kubenswrapper[4775]: I0321 05:08:46.320619 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="71255c4f-1e47-4e35-845f-876fff5fd6d4" containerName="glance-httpd" containerID="cri-o://9f447f1b0e6151bf34a0eec832692e899e0a001912e2cf06411e6494eecfb405" gracePeriod=30 Mar 21 05:08:46 crc kubenswrapper[4775]: I0321 05:08:46.391961 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:08:47 crc kubenswrapper[4775]: I0321 05:08:47.156038 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab91b8d2-32c9-4542-a730-7fc84e9892bd","Type":"ContainerStarted","Data":"756b7ac25394ff7e03c954f4dded0675f6295888345b0b7a098c3ad32426a8ef"} Mar 21 05:08:47 crc kubenswrapper[4775]: I0321 05:08:47.162021 4775 generic.go:334] "Generic (PLEG): container finished" podID="71255c4f-1e47-4e35-845f-876fff5fd6d4" containerID="b0bf1b828206099564568160f6163284ee71379761dd593fcfafe3139a339268" exitCode=143 Mar 21 05:08:47 crc kubenswrapper[4775]: I0321 05:08:47.162061 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"71255c4f-1e47-4e35-845f-876fff5fd6d4","Type":"ContainerDied","Data":"b0bf1b828206099564568160f6163284ee71379761dd593fcfafe3139a339268"} Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.184864 4775 generic.go:334] "Generic (PLEG): container finished" podID="47e577d7-e389-4135-b4fb-979bd627eaa9" containerID="0173c3dcb2f1ee362b645abaeead4a322f137b6409341a3b746ad5f19cffafb5" exitCode=0 Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.185150 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47e577d7-e389-4135-b4fb-979bd627eaa9","Type":"ContainerDied","Data":"0173c3dcb2f1ee362b645abaeead4a322f137b6409341a3b746ad5f19cffafb5"} Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.422977 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.582065 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-config-data\") pod \"47e577d7-e389-4135-b4fb-979bd627eaa9\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.582430 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4spw\" (UniqueName: \"kubernetes.io/projected/47e577d7-e389-4135-b4fb-979bd627eaa9-kube-api-access-d4spw\") pod \"47e577d7-e389-4135-b4fb-979bd627eaa9\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.582476 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-scripts\") pod \"47e577d7-e389-4135-b4fb-979bd627eaa9\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.582496 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/47e577d7-e389-4135-b4fb-979bd627eaa9-httpd-run\") pod \"47e577d7-e389-4135-b4fb-979bd627eaa9\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.582574 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-combined-ca-bundle\") pod \"47e577d7-e389-4135-b4fb-979bd627eaa9\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.582616 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"47e577d7-e389-4135-b4fb-979bd627eaa9\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.582635 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-public-tls-certs\") pod \"47e577d7-e389-4135-b4fb-979bd627eaa9\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.582670 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47e577d7-e389-4135-b4fb-979bd627eaa9-logs\") pod \"47e577d7-e389-4135-b4fb-979bd627eaa9\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.583303 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47e577d7-e389-4135-b4fb-979bd627eaa9-logs" (OuterVolumeSpecName: "logs") pod "47e577d7-e389-4135-b4fb-979bd627eaa9" (UID: "47e577d7-e389-4135-b4fb-979bd627eaa9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.583645 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47e577d7-e389-4135-b4fb-979bd627eaa9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "47e577d7-e389-4135-b4fb-979bd627eaa9" (UID: "47e577d7-e389-4135-b4fb-979bd627eaa9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.589307 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-scripts" (OuterVolumeSpecName: "scripts") pod "47e577d7-e389-4135-b4fb-979bd627eaa9" (UID: "47e577d7-e389-4135-b4fb-979bd627eaa9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.589329 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47e577d7-e389-4135-b4fb-979bd627eaa9-kube-api-access-d4spw" (OuterVolumeSpecName: "kube-api-access-d4spw") pod "47e577d7-e389-4135-b4fb-979bd627eaa9" (UID: "47e577d7-e389-4135-b4fb-979bd627eaa9"). InnerVolumeSpecName "kube-api-access-d4spw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.590354 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "47e577d7-e389-4135-b4fb-979bd627eaa9" (UID: "47e577d7-e389-4135-b4fb-979bd627eaa9"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.625543 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47e577d7-e389-4135-b4fb-979bd627eaa9" (UID: "47e577d7-e389-4135-b4fb-979bd627eaa9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:48 crc kubenswrapper[4775]: E0321 05:08:48.647713 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-public-tls-certs podName:47e577d7-e389-4135-b4fb-979bd627eaa9 nodeName:}" failed. No retries permitted until 2026-03-21 05:08:49.147680543 +0000 UTC m=+1282.124144167 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "public-tls-certs" (UniqueName: "kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-public-tls-certs") pod "47e577d7-e389-4135-b4fb-979bd627eaa9" (UID: "47e577d7-e389-4135-b4fb-979bd627eaa9") : error deleting /var/lib/kubelet/pods/47e577d7-e389-4135-b4fb-979bd627eaa9/volume-subpaths: remove /var/lib/kubelet/pods/47e577d7-e389-4135-b4fb-979bd627eaa9/volume-subpaths: no such file or directory Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.650780 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-config-data" (OuterVolumeSpecName: "config-data") pod "47e577d7-e389-4135-b4fb-979bd627eaa9" (UID: "47e577d7-e389-4135-b4fb-979bd627eaa9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.684812 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.685145 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.685255 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47e577d7-e389-4135-b4fb-979bd627eaa9-logs\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.685346 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.685432 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4spw\" (UniqueName: \"kubernetes.io/projected/47e577d7-e389-4135-b4fb-979bd627eaa9-kube-api-access-d4spw\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.685520 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.685590 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/47e577d7-e389-4135-b4fb-979bd627eaa9-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.720308 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Mar 21 05:08:48 crc kubenswrapper[4775]: I0321 05:08:48.787714 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.193265 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-public-tls-certs\") pod \"47e577d7-e389-4135-b4fb-979bd627eaa9\" (UID: \"47e577d7-e389-4135-b4fb-979bd627eaa9\") " Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.196878 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47e577d7-e389-4135-b4fb-979bd627eaa9","Type":"ContainerDied","Data":"ea4d39e040f38b33347813ecac429d1646c7747cc95596bc69a68249c6432f4a"} Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.197033 4775 scope.go:117] "RemoveContainer" containerID="0173c3dcb2f1ee362b645abaeead4a322f137b6409341a3b746ad5f19cffafb5" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.197081 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.200914 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab91b8d2-32c9-4542-a730-7fc84e9892bd","Type":"ContainerStarted","Data":"30455e621655fa15c45c8fad2a223bc59355675a9f5b3da03f27d5155ab96f2d"} Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.201188 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.201103 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerName="proxy-httpd" containerID="cri-o://30455e621655fa15c45c8fad2a223bc59355675a9f5b3da03f27d5155ab96f2d" gracePeriod=30 Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.201134 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerName="sg-core" containerID="cri-o://756b7ac25394ff7e03c954f4dded0675f6295888345b0b7a098c3ad32426a8ef" gracePeriod=30 Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.201160 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerName="ceilometer-notification-agent" containerID="cri-o://8cc8d28938a70addf2d6412cf322bdf56956d1adec39274fbeadbaacd685266d" gracePeriod=30 Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.200942 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerName="ceilometer-central-agent" containerID="cri-o://f3b6a0d856af07fa3386647a1df14c43107fa18d16407653836fe8f6805030db" gracePeriod=30 Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.203733 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "47e577d7-e389-4135-b4fb-979bd627eaa9" (UID: "47e577d7-e389-4135-b4fb-979bd627eaa9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.271607 4775 scope.go:117] "RemoveContainer" containerID="29a18bd4308c28984ec8d5e5e490324db58708c1d07372b24ee7783e1995157d" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.297000 4775 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/47e577d7-e389-4135-b4fb-979bd627eaa9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.632021 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.690917812 podStartE2EDuration="6.63197881s" podCreationTimestamp="2026-03-21 05:08:43 +0000 UTC" firstStartedPulling="2026-03-21 05:08:43.967429776 +0000 UTC m=+1276.943893410" lastFinishedPulling="2026-03-21 05:08:48.908490794 +0000 UTC m=+1281.884954408" observedRunningTime="2026-03-21 05:08:49.225878173 +0000 UTC m=+1282.202341797" watchObservedRunningTime="2026-03-21 05:08:49.63197881 +0000 UTC m=+1282.608442434" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.640615 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.655861 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.673812 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47e577d7-e389-4135-b4fb-979bd627eaa9" path="/var/lib/kubelet/pods/47e577d7-e389-4135-b4fb-979bd627eaa9/volumes" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.691237 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 05:08:49 crc kubenswrapper[4775]: E0321 05:08:49.691797 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47e577d7-e389-4135-b4fb-979bd627eaa9" containerName="glance-httpd" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.691813 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="47e577d7-e389-4135-b4fb-979bd627eaa9" containerName="glance-httpd" Mar 21 05:08:49 crc kubenswrapper[4775]: E0321 05:08:49.691828 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47e577d7-e389-4135-b4fb-979bd627eaa9" containerName="glance-log" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.691835 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="47e577d7-e389-4135-b4fb-979bd627eaa9" containerName="glance-log" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.691996 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="47e577d7-e389-4135-b4fb-979bd627eaa9" containerName="glance-httpd" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.692026 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="47e577d7-e389-4135-b4fb-979bd627eaa9" containerName="glance-log" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.692943 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.700756 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.701763 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.702449 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.856027 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.856087 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01638b90-5e17-43b3-a3b5-90726b26e243-config-data\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.856151 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01638b90-5e17-43b3-a3b5-90726b26e243-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.856188 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/01638b90-5e17-43b3-a3b5-90726b26e243-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.856292 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01638b90-5e17-43b3-a3b5-90726b26e243-logs\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.856326 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01638b90-5e17-43b3-a3b5-90726b26e243-scripts\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.856354 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghqz9\" (UniqueName: \"kubernetes.io/projected/01638b90-5e17-43b3-a3b5-90726b26e243-kube-api-access-ghqz9\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.856404 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01638b90-5e17-43b3-a3b5-90726b26e243-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.958327 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01638b90-5e17-43b3-a3b5-90726b26e243-logs\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.958712 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01638b90-5e17-43b3-a3b5-90726b26e243-scripts\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.958752 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghqz9\" (UniqueName: \"kubernetes.io/projected/01638b90-5e17-43b3-a3b5-90726b26e243-kube-api-access-ghqz9\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.958796 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01638b90-5e17-43b3-a3b5-90726b26e243-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.958873 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.958911 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01638b90-5e17-43b3-a3b5-90726b26e243-config-data\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.958953 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01638b90-5e17-43b3-a3b5-90726b26e243-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.958988 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/01638b90-5e17-43b3-a3b5-90726b26e243-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.959887 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.959946 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/01638b90-5e17-43b3-a3b5-90726b26e243-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.959887 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01638b90-5e17-43b3-a3b5-90726b26e243-logs\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.975612 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01638b90-5e17-43b3-a3b5-90726b26e243-scripts\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.976613 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01638b90-5e17-43b3-a3b5-90726b26e243-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.977225 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01638b90-5e17-43b3-a3b5-90726b26e243-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:49 crc kubenswrapper[4775]: I0321 05:08:49.988174 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01638b90-5e17-43b3-a3b5-90726b26e243-config-data\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.008229 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghqz9\" (UniqueName: \"kubernetes.io/projected/01638b90-5e17-43b3-a3b5-90726b26e243-kube-api-access-ghqz9\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.013584 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"01638b90-5e17-43b3-a3b5-90726b26e243\") " pod="openstack/glance-default-external-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.022846 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.156796 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.257827 4775 generic.go:334] "Generic (PLEG): container finished" podID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerID="756b7ac25394ff7e03c954f4dded0675f6295888345b0b7a098c3ad32426a8ef" exitCode=2 Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.257855 4775 generic.go:334] "Generic (PLEG): container finished" podID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerID="8cc8d28938a70addf2d6412cf322bdf56956d1adec39274fbeadbaacd685266d" exitCode=0 Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.257906 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab91b8d2-32c9-4542-a730-7fc84e9892bd","Type":"ContainerDied","Data":"756b7ac25394ff7e03c954f4dded0675f6295888345b0b7a098c3ad32426a8ef"} Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.257934 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab91b8d2-32c9-4542-a730-7fc84e9892bd","Type":"ContainerDied","Data":"8cc8d28938a70addf2d6412cf322bdf56956d1adec39274fbeadbaacd685266d"} Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.260034 4775 generic.go:334] "Generic (PLEG): container finished" podID="71255c4f-1e47-4e35-845f-876fff5fd6d4" containerID="9f447f1b0e6151bf34a0eec832692e899e0a001912e2cf06411e6494eecfb405" exitCode=0 Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.260061 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"71255c4f-1e47-4e35-845f-876fff5fd6d4","Type":"ContainerDied","Data":"9f447f1b0e6151bf34a0eec832692e899e0a001912e2cf06411e6494eecfb405"} Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.260082 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"71255c4f-1e47-4e35-845f-876fff5fd6d4","Type":"ContainerDied","Data":"5d037f60a7382103283fb7dc977007a3668afac9d065f14a09fa2c15aa21b79e"} Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.260098 4775 scope.go:117] "RemoveContainer" containerID="9f447f1b0e6151bf34a0eec832692e899e0a001912e2cf06411e6494eecfb405" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.260266 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.266015 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-combined-ca-bundle\") pod \"71255c4f-1e47-4e35-845f-876fff5fd6d4\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.266061 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"71255c4f-1e47-4e35-845f-876fff5fd6d4\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.266079 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-internal-tls-certs\") pod \"71255c4f-1e47-4e35-845f-876fff5fd6d4\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.266169 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71255c4f-1e47-4e35-845f-876fff5fd6d4-logs\") pod \"71255c4f-1e47-4e35-845f-876fff5fd6d4\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.266203 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-config-data\") pod \"71255c4f-1e47-4e35-845f-876fff5fd6d4\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.266238 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-scripts\") pod \"71255c4f-1e47-4e35-845f-876fff5fd6d4\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.266291 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/71255c4f-1e47-4e35-845f-876fff5fd6d4-httpd-run\") pod \"71255c4f-1e47-4e35-845f-876fff5fd6d4\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.266365 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pz2pk\" (UniqueName: \"kubernetes.io/projected/71255c4f-1e47-4e35-845f-876fff5fd6d4-kube-api-access-pz2pk\") pod \"71255c4f-1e47-4e35-845f-876fff5fd6d4\" (UID: \"71255c4f-1e47-4e35-845f-876fff5fd6d4\") " Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.270881 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71255c4f-1e47-4e35-845f-876fff5fd6d4-logs" (OuterVolumeSpecName: "logs") pod "71255c4f-1e47-4e35-845f-876fff5fd6d4" (UID: "71255c4f-1e47-4e35-845f-876fff5fd6d4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.270901 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71255c4f-1e47-4e35-845f-876fff5fd6d4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "71255c4f-1e47-4e35-845f-876fff5fd6d4" (UID: "71255c4f-1e47-4e35-845f-876fff5fd6d4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.286960 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "71255c4f-1e47-4e35-845f-876fff5fd6d4" (UID: "71255c4f-1e47-4e35-845f-876fff5fd6d4"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.287028 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-scripts" (OuterVolumeSpecName: "scripts") pod "71255c4f-1e47-4e35-845f-876fff5fd6d4" (UID: "71255c4f-1e47-4e35-845f-876fff5fd6d4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.288100 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71255c4f-1e47-4e35-845f-876fff5fd6d4-kube-api-access-pz2pk" (OuterVolumeSpecName: "kube-api-access-pz2pk") pod "71255c4f-1e47-4e35-845f-876fff5fd6d4" (UID: "71255c4f-1e47-4e35-845f-876fff5fd6d4"). InnerVolumeSpecName "kube-api-access-pz2pk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.321396 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "71255c4f-1e47-4e35-845f-876fff5fd6d4" (UID: "71255c4f-1e47-4e35-845f-876fff5fd6d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.332250 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-config-data" (OuterVolumeSpecName: "config-data") pod "71255c4f-1e47-4e35-845f-876fff5fd6d4" (UID: "71255c4f-1e47-4e35-845f-876fff5fd6d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.364473 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "71255c4f-1e47-4e35-845f-876fff5fd6d4" (UID: "71255c4f-1e47-4e35-845f-876fff5fd6d4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.368402 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/71255c4f-1e47-4e35-845f-876fff5fd6d4-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.368439 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pz2pk\" (UniqueName: \"kubernetes.io/projected/71255c4f-1e47-4e35-845f-876fff5fd6d4-kube-api-access-pz2pk\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.368454 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.368466 4775 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.368504 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.368517 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71255c4f-1e47-4e35-845f-876fff5fd6d4-logs\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.368528 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.368539 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71255c4f-1e47-4e35-845f-876fff5fd6d4-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.397991 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.398418 4775 scope.go:117] "RemoveContainer" containerID="b0bf1b828206099564568160f6163284ee71379761dd593fcfafe3139a339268" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.438452 4775 scope.go:117] "RemoveContainer" containerID="9f447f1b0e6151bf34a0eec832692e899e0a001912e2cf06411e6494eecfb405" Mar 21 05:08:50 crc kubenswrapper[4775]: E0321 05:08:50.440943 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f447f1b0e6151bf34a0eec832692e899e0a001912e2cf06411e6494eecfb405\": container with ID starting with 9f447f1b0e6151bf34a0eec832692e899e0a001912e2cf06411e6494eecfb405 not found: ID does not exist" containerID="9f447f1b0e6151bf34a0eec832692e899e0a001912e2cf06411e6494eecfb405" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.441001 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f447f1b0e6151bf34a0eec832692e899e0a001912e2cf06411e6494eecfb405"} err="failed to get container status \"9f447f1b0e6151bf34a0eec832692e899e0a001912e2cf06411e6494eecfb405\": rpc error: code = NotFound desc = could not find container \"9f447f1b0e6151bf34a0eec832692e899e0a001912e2cf06411e6494eecfb405\": container with ID starting with 9f447f1b0e6151bf34a0eec832692e899e0a001912e2cf06411e6494eecfb405 not found: ID does not exist" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.441035 4775 scope.go:117] "RemoveContainer" containerID="b0bf1b828206099564568160f6163284ee71379761dd593fcfafe3139a339268" Mar 21 05:08:50 crc kubenswrapper[4775]: E0321 05:08:50.445572 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0bf1b828206099564568160f6163284ee71379761dd593fcfafe3139a339268\": container with ID starting with b0bf1b828206099564568160f6163284ee71379761dd593fcfafe3139a339268 not found: ID does not exist" containerID="b0bf1b828206099564568160f6163284ee71379761dd593fcfafe3139a339268" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.445655 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0bf1b828206099564568160f6163284ee71379761dd593fcfafe3139a339268"} err="failed to get container status \"b0bf1b828206099564568160f6163284ee71379761dd593fcfafe3139a339268\": rpc error: code = NotFound desc = could not find container \"b0bf1b828206099564568160f6163284ee71379761dd593fcfafe3139a339268\": container with ID starting with b0bf1b828206099564568160f6163284ee71379761dd593fcfafe3139a339268 not found: ID does not exist" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.469967 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.607413 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.618390 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.644274 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 05:08:50 crc kubenswrapper[4775]: E0321 05:08:50.644697 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71255c4f-1e47-4e35-845f-876fff5fd6d4" containerName="glance-httpd" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.644710 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="71255c4f-1e47-4e35-845f-876fff5fd6d4" containerName="glance-httpd" Mar 21 05:08:50 crc kubenswrapper[4775]: E0321 05:08:50.644725 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71255c4f-1e47-4e35-845f-876fff5fd6d4" containerName="glance-log" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.644732 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="71255c4f-1e47-4e35-845f-876fff5fd6d4" containerName="glance-log" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.644934 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="71255c4f-1e47-4e35-845f-876fff5fd6d4" containerName="glance-log" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.644958 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="71255c4f-1e47-4e35-845f-876fff5fd6d4" containerName="glance-httpd" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.646068 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.649252 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.649539 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.667573 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.721951 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.776733 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.776880 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-logs\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.776957 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.777109 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.777186 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.777220 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.777377 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftppj\" (UniqueName: \"kubernetes.io/projected/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-kube-api-access-ftppj\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.777670 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.879407 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.879460 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-logs\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.879493 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.879545 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.879569 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.879591 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.879639 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftppj\" (UniqueName: \"kubernetes.io/projected/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-kube-api-access-ftppj\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.879797 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.879789 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.880216 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-logs\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.880265 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.884671 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.884702 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.885783 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.894543 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.902474 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftppj\" (UniqueName: \"kubernetes.io/projected/e9e72c4b-a3fb-41eb-974a-74d24d6cdac9-kube-api-access-ftppj\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.911051 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9\") " pod="openstack/glance-default-internal-api-0" Mar 21 05:08:50 crc kubenswrapper[4775]: I0321 05:08:50.964839 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 05:08:51 crc kubenswrapper[4775]: I0321 05:08:51.271254 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"01638b90-5e17-43b3-a3b5-90726b26e243","Type":"ContainerStarted","Data":"306032f21837c9eaad68cb251fcdd340b028c66fc621cccbb829910f80e9b9eb"} Mar 21 05:08:51 crc kubenswrapper[4775]: I0321 05:08:51.561170 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 05:08:51 crc kubenswrapper[4775]: W0321 05:08:51.573226 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9e72c4b_a3fb_41eb_974a_74d24d6cdac9.slice/crio-43fe5d0cb79c2f9695f700da1f42f8761f2efb036bd157b8a381ccb30c2c5b5e WatchSource:0}: Error finding container 43fe5d0cb79c2f9695f700da1f42f8761f2efb036bd157b8a381ccb30c2c5b5e: Status 404 returned error can't find the container with id 43fe5d0cb79c2f9695f700da1f42f8761f2efb036bd157b8a381ccb30c2c5b5e Mar 21 05:08:51 crc kubenswrapper[4775]: I0321 05:08:51.678148 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71255c4f-1e47-4e35-845f-876fff5fd6d4" path="/var/lib/kubelet/pods/71255c4f-1e47-4e35-845f-876fff5fd6d4/volumes" Mar 21 05:08:52 crc kubenswrapper[4775]: I0321 05:08:52.293759 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9","Type":"ContainerStarted","Data":"8c18a8a69790fe1df7223ccfc07aa1d02e91b4c205aeb964f24e898e96a99bd7"} Mar 21 05:08:52 crc kubenswrapper[4775]: I0321 05:08:52.294043 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9","Type":"ContainerStarted","Data":"43fe5d0cb79c2f9695f700da1f42f8761f2efb036bd157b8a381ccb30c2c5b5e"} Mar 21 05:08:52 crc kubenswrapper[4775]: I0321 05:08:52.297391 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"01638b90-5e17-43b3-a3b5-90726b26e243","Type":"ContainerStarted","Data":"bb11d5c8afbf3756020e0b603783b8b46872e5fea4262f38b6c80e6d01e2ae3c"} Mar 21 05:08:52 crc kubenswrapper[4775]: I0321 05:08:52.297437 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"01638b90-5e17-43b3-a3b5-90726b26e243","Type":"ContainerStarted","Data":"e388a0ecedcb04f4424057bc2c07d6e3f4d796b885be7531b5e2d7a5759e1c9b"} Mar 21 05:08:52 crc kubenswrapper[4775]: I0321 05:08:52.326991 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.326967842 podStartE2EDuration="3.326967842s" podCreationTimestamp="2026-03-21 05:08:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:08:52.315447776 +0000 UTC m=+1285.291911400" watchObservedRunningTime="2026-03-21 05:08:52.326967842 +0000 UTC m=+1285.303431476" Mar 21 05:08:53 crc kubenswrapper[4775]: I0321 05:08:53.311228 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e9e72c4b-a3fb-41eb-974a-74d24d6cdac9","Type":"ContainerStarted","Data":"902eb75dc6f362171ab35925bd1c393028b1b3fbaa2d60e5676d0154648ce94f"} Mar 21 05:08:53 crc kubenswrapper[4775]: I0321 05:08:53.341160 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.341131693 podStartE2EDuration="3.341131693s" podCreationTimestamp="2026-03-21 05:08:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:08:53.33183995 +0000 UTC m=+1286.308303574" watchObservedRunningTime="2026-03-21 05:08:53.341131693 +0000 UTC m=+1286.317595317" Mar 21 05:08:54 crc kubenswrapper[4775]: I0321 05:08:54.322057 4775 generic.go:334] "Generic (PLEG): container finished" podID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerID="f3b6a0d856af07fa3386647a1df14c43107fa18d16407653836fe8f6805030db" exitCode=0 Mar 21 05:08:54 crc kubenswrapper[4775]: I0321 05:08:54.322111 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab91b8d2-32c9-4542-a730-7fc84e9892bd","Type":"ContainerDied","Data":"f3b6a0d856af07fa3386647a1df14c43107fa18d16407653836fe8f6805030db"} Mar 21 05:08:54 crc kubenswrapper[4775]: I0321 05:08:54.323887 4775 generic.go:334] "Generic (PLEG): container finished" podID="824625b1-30cc-42c5-ad83-5854770c2f46" containerID="20d707fa7f70310d9cb80b150c04d331c2047a421ec95e9ba298af9cad00d19d" exitCode=0 Mar 21 05:08:54 crc kubenswrapper[4775]: I0321 05:08:54.323919 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gk78h" event={"ID":"824625b1-30cc-42c5-ad83-5854770c2f46","Type":"ContainerDied","Data":"20d707fa7f70310d9cb80b150c04d331c2047a421ec95e9ba298af9cad00d19d"} Mar 21 05:08:55 crc kubenswrapper[4775]: I0321 05:08:55.702907 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gk78h" Mar 21 05:08:55 crc kubenswrapper[4775]: I0321 05:08:55.767994 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/824625b1-30cc-42c5-ad83-5854770c2f46-config-data\") pod \"824625b1-30cc-42c5-ad83-5854770c2f46\" (UID: \"824625b1-30cc-42c5-ad83-5854770c2f46\") " Mar 21 05:08:55 crc kubenswrapper[4775]: I0321 05:08:55.768192 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/824625b1-30cc-42c5-ad83-5854770c2f46-combined-ca-bundle\") pod \"824625b1-30cc-42c5-ad83-5854770c2f46\" (UID: \"824625b1-30cc-42c5-ad83-5854770c2f46\") " Mar 21 05:08:55 crc kubenswrapper[4775]: I0321 05:08:55.768304 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/824625b1-30cc-42c5-ad83-5854770c2f46-scripts\") pod \"824625b1-30cc-42c5-ad83-5854770c2f46\" (UID: \"824625b1-30cc-42c5-ad83-5854770c2f46\") " Mar 21 05:08:55 crc kubenswrapper[4775]: I0321 05:08:55.768338 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59jrc\" (UniqueName: \"kubernetes.io/projected/824625b1-30cc-42c5-ad83-5854770c2f46-kube-api-access-59jrc\") pod \"824625b1-30cc-42c5-ad83-5854770c2f46\" (UID: \"824625b1-30cc-42c5-ad83-5854770c2f46\") " Mar 21 05:08:55 crc kubenswrapper[4775]: I0321 05:08:55.774268 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/824625b1-30cc-42c5-ad83-5854770c2f46-scripts" (OuterVolumeSpecName: "scripts") pod "824625b1-30cc-42c5-ad83-5854770c2f46" (UID: "824625b1-30cc-42c5-ad83-5854770c2f46"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:55 crc kubenswrapper[4775]: I0321 05:08:55.774296 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/824625b1-30cc-42c5-ad83-5854770c2f46-kube-api-access-59jrc" (OuterVolumeSpecName: "kube-api-access-59jrc") pod "824625b1-30cc-42c5-ad83-5854770c2f46" (UID: "824625b1-30cc-42c5-ad83-5854770c2f46"). InnerVolumeSpecName "kube-api-access-59jrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:08:55 crc kubenswrapper[4775]: I0321 05:08:55.809101 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/824625b1-30cc-42c5-ad83-5854770c2f46-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "824625b1-30cc-42c5-ad83-5854770c2f46" (UID: "824625b1-30cc-42c5-ad83-5854770c2f46"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:55 crc kubenswrapper[4775]: I0321 05:08:55.820992 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/824625b1-30cc-42c5-ad83-5854770c2f46-config-data" (OuterVolumeSpecName: "config-data") pod "824625b1-30cc-42c5-ad83-5854770c2f46" (UID: "824625b1-30cc-42c5-ad83-5854770c2f46"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:08:55 crc kubenswrapper[4775]: I0321 05:08:55.870894 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/824625b1-30cc-42c5-ad83-5854770c2f46-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:55 crc kubenswrapper[4775]: I0321 05:08:55.870945 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/824625b1-30cc-42c5-ad83-5854770c2f46-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:55 crc kubenswrapper[4775]: I0321 05:08:55.870961 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59jrc\" (UniqueName: \"kubernetes.io/projected/824625b1-30cc-42c5-ad83-5854770c2f46-kube-api-access-59jrc\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:55 crc kubenswrapper[4775]: I0321 05:08:55.870978 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/824625b1-30cc-42c5-ad83-5854770c2f46-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:08:56 crc kubenswrapper[4775]: I0321 05:08:56.347622 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gk78h" event={"ID":"824625b1-30cc-42c5-ad83-5854770c2f46","Type":"ContainerDied","Data":"2d77ca12f9d9586582c849b2cbdaffac8b8dcfcc0b7a10cdf251cde2b34e8f90"} Mar 21 05:08:56 crc kubenswrapper[4775]: I0321 05:08:56.347661 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d77ca12f9d9586582c849b2cbdaffac8b8dcfcc0b7a10cdf251cde2b34e8f90" Mar 21 05:08:56 crc kubenswrapper[4775]: I0321 05:08:56.347669 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gk78h" Mar 21 05:08:56 crc kubenswrapper[4775]: I0321 05:08:56.448967 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 21 05:08:56 crc kubenswrapper[4775]: E0321 05:08:56.449513 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="824625b1-30cc-42c5-ad83-5854770c2f46" containerName="nova-cell0-conductor-db-sync" Mar 21 05:08:56 crc kubenswrapper[4775]: I0321 05:08:56.449537 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="824625b1-30cc-42c5-ad83-5854770c2f46" containerName="nova-cell0-conductor-db-sync" Mar 21 05:08:56 crc kubenswrapper[4775]: I0321 05:08:56.449764 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="824625b1-30cc-42c5-ad83-5854770c2f46" containerName="nova-cell0-conductor-db-sync" Mar 21 05:08:56 crc kubenswrapper[4775]: I0321 05:08:56.450504 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 21 05:08:56 crc kubenswrapper[4775]: I0321 05:08:56.455853 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 21 05:08:56 crc kubenswrapper[4775]: I0321 05:08:56.456111 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bhr6z" Mar 21 05:08:56 crc kubenswrapper[4775]: I0321 05:08:56.467278 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 21 05:08:56 crc kubenswrapper[4775]: I0321 05:08:56.483933 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nl54\" (UniqueName: \"kubernetes.io/projected/789a0bb8-b131-4144-9400-7c32a604d6d5-kube-api-access-5nl54\") pod \"nova-cell0-conductor-0\" (UID: \"789a0bb8-b131-4144-9400-7c32a604d6d5\") " pod="openstack/nova-cell0-conductor-0" Mar 21 05:08:56 crc kubenswrapper[4775]: I0321 05:08:56.484021 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/789a0bb8-b131-4144-9400-7c32a604d6d5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"789a0bb8-b131-4144-9400-7c32a604d6d5\") " pod="openstack/nova-cell0-conductor-0" Mar 21 05:08:56 crc kubenswrapper[4775]: I0321 05:08:56.484164 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/789a0bb8-b131-4144-9400-7c32a604d6d5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"789a0bb8-b131-4144-9400-7c32a604d6d5\") " pod="openstack/nova-cell0-conductor-0" Mar 21 05:08:56 crc kubenswrapper[4775]: I0321 05:08:56.585586 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nl54\" (UniqueName: \"kubernetes.io/projected/789a0bb8-b131-4144-9400-7c32a604d6d5-kube-api-access-5nl54\") pod \"nova-cell0-conductor-0\" (UID: \"789a0bb8-b131-4144-9400-7c32a604d6d5\") " pod="openstack/nova-cell0-conductor-0" Mar 21 05:08:56 crc kubenswrapper[4775]: I0321 05:08:56.585680 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/789a0bb8-b131-4144-9400-7c32a604d6d5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"789a0bb8-b131-4144-9400-7c32a604d6d5\") " pod="openstack/nova-cell0-conductor-0" Mar 21 05:08:56 crc kubenswrapper[4775]: I0321 05:08:56.585774 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/789a0bb8-b131-4144-9400-7c32a604d6d5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"789a0bb8-b131-4144-9400-7c32a604d6d5\") " pod="openstack/nova-cell0-conductor-0" Mar 21 05:08:56 crc kubenswrapper[4775]: I0321 05:08:56.590576 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/789a0bb8-b131-4144-9400-7c32a604d6d5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"789a0bb8-b131-4144-9400-7c32a604d6d5\") " pod="openstack/nova-cell0-conductor-0" Mar 21 05:08:56 crc kubenswrapper[4775]: I0321 05:08:56.591029 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/789a0bb8-b131-4144-9400-7c32a604d6d5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"789a0bb8-b131-4144-9400-7c32a604d6d5\") " pod="openstack/nova-cell0-conductor-0" Mar 21 05:08:56 crc kubenswrapper[4775]: I0321 05:08:56.604393 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nl54\" (UniqueName: \"kubernetes.io/projected/789a0bb8-b131-4144-9400-7c32a604d6d5-kube-api-access-5nl54\") pod \"nova-cell0-conductor-0\" (UID: \"789a0bb8-b131-4144-9400-7c32a604d6d5\") " pod="openstack/nova-cell0-conductor-0" Mar 21 05:08:56 crc kubenswrapper[4775]: I0321 05:08:56.775339 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 21 05:08:57 crc kubenswrapper[4775]: I0321 05:08:57.285135 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 21 05:08:57 crc kubenswrapper[4775]: I0321 05:08:57.356343 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"789a0bb8-b131-4144-9400-7c32a604d6d5","Type":"ContainerStarted","Data":"55e62127f17f2e0706ae755a8c32b956ddd17f11e1e5e7eab7ea44e3cbd3b157"} Mar 21 05:08:58 crc kubenswrapper[4775]: I0321 05:08:58.365838 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"789a0bb8-b131-4144-9400-7c32a604d6d5","Type":"ContainerStarted","Data":"b784379bb60e5f18b33c1593dc0cd9eefd62b1907352006d148b3c310fb65252"} Mar 21 05:08:58 crc kubenswrapper[4775]: I0321 05:08:58.367361 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 21 05:08:58 crc kubenswrapper[4775]: I0321 05:08:58.396370 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.396351336 podStartE2EDuration="2.396351336s" podCreationTimestamp="2026-03-21 05:08:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:08:58.39155521 +0000 UTC m=+1291.368018844" watchObservedRunningTime="2026-03-21 05:08:58.396351336 +0000 UTC m=+1291.372814970" Mar 21 05:09:00 crc kubenswrapper[4775]: I0321 05:09:00.024464 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 21 05:09:00 crc kubenswrapper[4775]: I0321 05:09:00.024765 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 21 05:09:00 crc kubenswrapper[4775]: I0321 05:09:00.052256 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 21 05:09:00 crc kubenswrapper[4775]: I0321 05:09:00.067457 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 21 05:09:00 crc kubenswrapper[4775]: I0321 05:09:00.386741 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 21 05:09:00 crc kubenswrapper[4775]: I0321 05:09:00.386804 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 21 05:09:00 crc kubenswrapper[4775]: I0321 05:09:00.965880 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 21 05:09:00 crc kubenswrapper[4775]: I0321 05:09:00.965942 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 21 05:09:00 crc kubenswrapper[4775]: I0321 05:09:00.995632 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 21 05:09:01 crc kubenswrapper[4775]: I0321 05:09:01.018208 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 21 05:09:01 crc kubenswrapper[4775]: I0321 05:09:01.395070 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 21 05:09:01 crc kubenswrapper[4775]: I0321 05:09:01.395104 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 21 05:09:02 crc kubenswrapper[4775]: I0321 05:09:02.312364 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 21 05:09:02 crc kubenswrapper[4775]: I0321 05:09:02.321032 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 21 05:09:03 crc kubenswrapper[4775]: I0321 05:09:03.345712 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 21 05:09:03 crc kubenswrapper[4775]: I0321 05:09:03.382695 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 21 05:09:06 crc kubenswrapper[4775]: I0321 05:09:06.803770 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.281219 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-pckp8"] Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.282601 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-pckp8" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.284601 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.284862 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.302286 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-pckp8"] Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.402545 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbdhj\" (UniqueName: \"kubernetes.io/projected/5a407dd6-593f-4806-884b-8e031639d25d-kube-api-access-zbdhj\") pod \"nova-cell0-cell-mapping-pckp8\" (UID: \"5a407dd6-593f-4806-884b-8e031639d25d\") " pod="openstack/nova-cell0-cell-mapping-pckp8" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.402664 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a407dd6-593f-4806-884b-8e031639d25d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-pckp8\" (UID: \"5a407dd6-593f-4806-884b-8e031639d25d\") " pod="openstack/nova-cell0-cell-mapping-pckp8" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.402818 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a407dd6-593f-4806-884b-8e031639d25d-scripts\") pod \"nova-cell0-cell-mapping-pckp8\" (UID: \"5a407dd6-593f-4806-884b-8e031639d25d\") " pod="openstack/nova-cell0-cell-mapping-pckp8" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.402911 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a407dd6-593f-4806-884b-8e031639d25d-config-data\") pod \"nova-cell0-cell-mapping-pckp8\" (UID: \"5a407dd6-593f-4806-884b-8e031639d25d\") " pod="openstack/nova-cell0-cell-mapping-pckp8" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.504688 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a407dd6-593f-4806-884b-8e031639d25d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-pckp8\" (UID: \"5a407dd6-593f-4806-884b-8e031639d25d\") " pod="openstack/nova-cell0-cell-mapping-pckp8" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.505058 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a407dd6-593f-4806-884b-8e031639d25d-scripts\") pod \"nova-cell0-cell-mapping-pckp8\" (UID: \"5a407dd6-593f-4806-884b-8e031639d25d\") " pod="openstack/nova-cell0-cell-mapping-pckp8" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.505217 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a407dd6-593f-4806-884b-8e031639d25d-config-data\") pod \"nova-cell0-cell-mapping-pckp8\" (UID: \"5a407dd6-593f-4806-884b-8e031639d25d\") " pod="openstack/nova-cell0-cell-mapping-pckp8" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.505345 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbdhj\" (UniqueName: \"kubernetes.io/projected/5a407dd6-593f-4806-884b-8e031639d25d-kube-api-access-zbdhj\") pod \"nova-cell0-cell-mapping-pckp8\" (UID: \"5a407dd6-593f-4806-884b-8e031639d25d\") " pod="openstack/nova-cell0-cell-mapping-pckp8" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.515817 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a407dd6-593f-4806-884b-8e031639d25d-scripts\") pod \"nova-cell0-cell-mapping-pckp8\" (UID: \"5a407dd6-593f-4806-884b-8e031639d25d\") " pod="openstack/nova-cell0-cell-mapping-pckp8" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.527190 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.528753 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.533455 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a407dd6-593f-4806-884b-8e031639d25d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-pckp8\" (UID: \"5a407dd6-593f-4806-884b-8e031639d25d\") " pod="openstack/nova-cell0-cell-mapping-pckp8" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.534230 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a407dd6-593f-4806-884b-8e031639d25d-config-data\") pod \"nova-cell0-cell-mapping-pckp8\" (UID: \"5a407dd6-593f-4806-884b-8e031639d25d\") " pod="openstack/nova-cell0-cell-mapping-pckp8" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.536107 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.555784 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbdhj\" (UniqueName: \"kubernetes.io/projected/5a407dd6-593f-4806-884b-8e031639d25d-kube-api-access-zbdhj\") pod \"nova-cell0-cell-mapping-pckp8\" (UID: \"5a407dd6-593f-4806-884b-8e031639d25d\") " pod="openstack/nova-cell0-cell-mapping-pckp8" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.570774 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.595314 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.599193 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.605762 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-pckp8" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.606765 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.618980 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.698445 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.699867 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.705627 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.715542 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87dad904-a84b-4dcf-9caa-3fe89053b96b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"87dad904-a84b-4dcf-9caa-3fe89053b96b\") " pod="openstack/nova-api-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.715630 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7219652-7280-4a81-b501-b6e41673162b-logs\") pod \"nova-metadata-0\" (UID: \"d7219652-7280-4a81-b501-b6e41673162b\") " pod="openstack/nova-metadata-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.715684 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9m8c\" (UniqueName: \"kubernetes.io/projected/d7219652-7280-4a81-b501-b6e41673162b-kube-api-access-g9m8c\") pod \"nova-metadata-0\" (UID: \"d7219652-7280-4a81-b501-b6e41673162b\") " pod="openstack/nova-metadata-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.715711 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7219652-7280-4a81-b501-b6e41673162b-config-data\") pod \"nova-metadata-0\" (UID: \"d7219652-7280-4a81-b501-b6e41673162b\") " pod="openstack/nova-metadata-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.715744 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7219652-7280-4a81-b501-b6e41673162b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d7219652-7280-4a81-b501-b6e41673162b\") " pod="openstack/nova-metadata-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.715768 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87dad904-a84b-4dcf-9caa-3fe89053b96b-config-data\") pod \"nova-api-0\" (UID: \"87dad904-a84b-4dcf-9caa-3fe89053b96b\") " pod="openstack/nova-api-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.715841 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e114fa87-9e65-4a0b-bad1-81f521aeef85-config-data\") pod \"nova-scheduler-0\" (UID: \"e114fa87-9e65-4a0b-bad1-81f521aeef85\") " pod="openstack/nova-scheduler-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.715884 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87dad904-a84b-4dcf-9caa-3fe89053b96b-logs\") pod \"nova-api-0\" (UID: \"87dad904-a84b-4dcf-9caa-3fe89053b96b\") " pod="openstack/nova-api-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.715923 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txpj5\" (UniqueName: \"kubernetes.io/projected/87dad904-a84b-4dcf-9caa-3fe89053b96b-kube-api-access-txpj5\") pod \"nova-api-0\" (UID: \"87dad904-a84b-4dcf-9caa-3fe89053b96b\") " pod="openstack/nova-api-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.716032 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpl5g\" (UniqueName: \"kubernetes.io/projected/e114fa87-9e65-4a0b-bad1-81f521aeef85-kube-api-access-xpl5g\") pod \"nova-scheduler-0\" (UID: \"e114fa87-9e65-4a0b-bad1-81f521aeef85\") " pod="openstack/nova-scheduler-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.716076 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e114fa87-9e65-4a0b-bad1-81f521aeef85-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e114fa87-9e65-4a0b-bad1-81f521aeef85\") " pod="openstack/nova-scheduler-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.730070 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.773870 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-qsbbk"] Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.775672 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.820283 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e114fa87-9e65-4a0b-bad1-81f521aeef85-config-data\") pod \"nova-scheduler-0\" (UID: \"e114fa87-9e65-4a0b-bad1-81f521aeef85\") " pod="openstack/nova-scheduler-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.820359 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87dad904-a84b-4dcf-9caa-3fe89053b96b-logs\") pod \"nova-api-0\" (UID: \"87dad904-a84b-4dcf-9caa-3fe89053b96b\") " pod="openstack/nova-api-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.820405 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptvg9\" (UniqueName: \"kubernetes.io/projected/20a97ec2-24be-494f-a3e8-dc7d202021c3-kube-api-access-ptvg9\") pod \"dnsmasq-dns-bccf8f775-qsbbk\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.820437 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txpj5\" (UniqueName: \"kubernetes.io/projected/87dad904-a84b-4dcf-9caa-3fe89053b96b-kube-api-access-txpj5\") pod \"nova-api-0\" (UID: \"87dad904-a84b-4dcf-9caa-3fe89053b96b\") " pod="openstack/nova-api-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.820506 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-dns-svc\") pod \"dnsmasq-dns-bccf8f775-qsbbk\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.820548 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpl5g\" (UniqueName: \"kubernetes.io/projected/e114fa87-9e65-4a0b-bad1-81f521aeef85-kube-api-access-xpl5g\") pod \"nova-scheduler-0\" (UID: \"e114fa87-9e65-4a0b-bad1-81f521aeef85\") " pod="openstack/nova-scheduler-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.820581 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e114fa87-9e65-4a0b-bad1-81f521aeef85-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e114fa87-9e65-4a0b-bad1-81f521aeef85\") " pod="openstack/nova-scheduler-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.820606 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-qsbbk\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.820637 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87dad904-a84b-4dcf-9caa-3fe89053b96b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"87dad904-a84b-4dcf-9caa-3fe89053b96b\") " pod="openstack/nova-api-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.820689 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-config\") pod \"dnsmasq-dns-bccf8f775-qsbbk\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.820715 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7219652-7280-4a81-b501-b6e41673162b-logs\") pod \"nova-metadata-0\" (UID: \"d7219652-7280-4a81-b501-b6e41673162b\") " pod="openstack/nova-metadata-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.820736 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-qsbbk\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.820761 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9m8c\" (UniqueName: \"kubernetes.io/projected/d7219652-7280-4a81-b501-b6e41673162b-kube-api-access-g9m8c\") pod \"nova-metadata-0\" (UID: \"d7219652-7280-4a81-b501-b6e41673162b\") " pod="openstack/nova-metadata-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.820788 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7219652-7280-4a81-b501-b6e41673162b-config-data\") pod \"nova-metadata-0\" (UID: \"d7219652-7280-4a81-b501-b6e41673162b\") " pod="openstack/nova-metadata-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.820820 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7219652-7280-4a81-b501-b6e41673162b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d7219652-7280-4a81-b501-b6e41673162b\") " pod="openstack/nova-metadata-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.820845 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87dad904-a84b-4dcf-9caa-3fe89053b96b-config-data\") pod \"nova-api-0\" (UID: \"87dad904-a84b-4dcf-9caa-3fe89053b96b\") " pod="openstack/nova-api-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.820873 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-qsbbk\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.823523 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7219652-7280-4a81-b501-b6e41673162b-logs\") pod \"nova-metadata-0\" (UID: \"d7219652-7280-4a81-b501-b6e41673162b\") " pod="openstack/nova-metadata-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.824406 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87dad904-a84b-4dcf-9caa-3fe89053b96b-logs\") pod \"nova-api-0\" (UID: \"87dad904-a84b-4dcf-9caa-3fe89053b96b\") " pod="openstack/nova-api-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.841845 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e114fa87-9e65-4a0b-bad1-81f521aeef85-config-data\") pod \"nova-scheduler-0\" (UID: \"e114fa87-9e65-4a0b-bad1-81f521aeef85\") " pod="openstack/nova-scheduler-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.841956 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87dad904-a84b-4dcf-9caa-3fe89053b96b-config-data\") pod \"nova-api-0\" (UID: \"87dad904-a84b-4dcf-9caa-3fe89053b96b\") " pod="openstack/nova-api-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.842832 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87dad904-a84b-4dcf-9caa-3fe89053b96b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"87dad904-a84b-4dcf-9caa-3fe89053b96b\") " pod="openstack/nova-api-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.843937 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7219652-7280-4a81-b501-b6e41673162b-config-data\") pod \"nova-metadata-0\" (UID: \"d7219652-7280-4a81-b501-b6e41673162b\") " pod="openstack/nova-metadata-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.849818 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpl5g\" (UniqueName: \"kubernetes.io/projected/e114fa87-9e65-4a0b-bad1-81f521aeef85-kube-api-access-xpl5g\") pod \"nova-scheduler-0\" (UID: \"e114fa87-9e65-4a0b-bad1-81f521aeef85\") " pod="openstack/nova-scheduler-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.851162 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7219652-7280-4a81-b501-b6e41673162b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d7219652-7280-4a81-b501-b6e41673162b\") " pod="openstack/nova-metadata-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.859378 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e114fa87-9e65-4a0b-bad1-81f521aeef85-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e114fa87-9e65-4a0b-bad1-81f521aeef85\") " pod="openstack/nova-scheduler-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.859376 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9m8c\" (UniqueName: \"kubernetes.io/projected/d7219652-7280-4a81-b501-b6e41673162b-kube-api-access-g9m8c\") pod \"nova-metadata-0\" (UID: \"d7219652-7280-4a81-b501-b6e41673162b\") " pod="openstack/nova-metadata-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.859947 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txpj5\" (UniqueName: \"kubernetes.io/projected/87dad904-a84b-4dcf-9caa-3fe89053b96b-kube-api-access-txpj5\") pod \"nova-api-0\" (UID: \"87dad904-a84b-4dcf-9caa-3fe89053b96b\") " pod="openstack/nova-api-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.878198 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-qsbbk"] Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.889217 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.890458 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.902328 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.923606 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.923677 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-dns-svc\") pod \"dnsmasq-dns-bccf8f775-qsbbk\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.923723 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swsh9\" (UniqueName: \"kubernetes.io/projected/e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2-kube-api-access-swsh9\") pod \"nova-cell1-novncproxy-0\" (UID: \"e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.923760 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-qsbbk\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.923806 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.923852 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-config\") pod \"dnsmasq-dns-bccf8f775-qsbbk\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.923883 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-qsbbk\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.923931 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-qsbbk\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.924002 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptvg9\" (UniqueName: \"kubernetes.io/projected/20a97ec2-24be-494f-a3e8-dc7d202021c3-kube-api-access-ptvg9\") pod \"dnsmasq-dns-bccf8f775-qsbbk\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.925519 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-dns-svc\") pod \"dnsmasq-dns-bccf8f775-qsbbk\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.927482 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-qsbbk\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.928361 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-qsbbk\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.931661 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-config\") pod \"dnsmasq-dns-bccf8f775-qsbbk\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.936497 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.937604 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-qsbbk\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.950421 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptvg9\" (UniqueName: \"kubernetes.io/projected/20a97ec2-24be-494f-a3e8-dc7d202021c3-kube-api-access-ptvg9\") pod \"dnsmasq-dns-bccf8f775-qsbbk\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:07 crc kubenswrapper[4775]: I0321 05:09:07.956060 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.026300 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.029889 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.030205 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swsh9\" (UniqueName: \"kubernetes.io/projected/e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2-kube-api-access-swsh9\") pod \"nova-cell1-novncproxy-0\" (UID: \"e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.030275 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.039104 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.045645 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.051916 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swsh9\" (UniqueName: \"kubernetes.io/projected/e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2-kube-api-access-swsh9\") pod \"nova-cell1-novncproxy-0\" (UID: \"e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.054406 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.102396 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.244813 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.314130 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-pckp8"] Mar 21 05:09:08 crc kubenswrapper[4775]: W0321 05:09:08.318076 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a407dd6_593f_4806_884b_8e031639d25d.slice/crio-040855e862320088965a0b53c32aefe654235d67e7c2aec1b7da6cf02854cd90 WatchSource:0}: Error finding container 040855e862320088965a0b53c32aefe654235d67e7c2aec1b7da6cf02854cd90: Status 404 returned error can't find the container with id 040855e862320088965a0b53c32aefe654235d67e7c2aec1b7da6cf02854cd90 Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.494183 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-pckp8" event={"ID":"5a407dd6-593f-4806-884b-8e031639d25d","Type":"ContainerStarted","Data":"040855e862320088965a0b53c32aefe654235d67e7c2aec1b7da6cf02854cd90"} Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.561424 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vd764"] Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.562475 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vd764" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.571499 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.571693 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.662816 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb1dd6e9-8801-464e-952c-d345498b132a-config-data\") pod \"nova-cell1-conductor-db-sync-vd764\" (UID: \"fb1dd6e9-8801-464e-952c-d345498b132a\") " pod="openstack/nova-cell1-conductor-db-sync-vd764" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.696321 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fmmn\" (UniqueName: \"kubernetes.io/projected/fb1dd6e9-8801-464e-952c-d345498b132a-kube-api-access-5fmmn\") pod \"nova-cell1-conductor-db-sync-vd764\" (UID: \"fb1dd6e9-8801-464e-952c-d345498b132a\") " pod="openstack/nova-cell1-conductor-db-sync-vd764" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.696607 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb1dd6e9-8801-464e-952c-d345498b132a-scripts\") pod \"nova-cell1-conductor-db-sync-vd764\" (UID: \"fb1dd6e9-8801-464e-952c-d345498b132a\") " pod="openstack/nova-cell1-conductor-db-sync-vd764" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.696733 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb1dd6e9-8801-464e-952c-d345498b132a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vd764\" (UID: \"fb1dd6e9-8801-464e-952c-d345498b132a\") " pod="openstack/nova-cell1-conductor-db-sync-vd764" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.743715 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vd764"] Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.810307 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb1dd6e9-8801-464e-952c-d345498b132a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vd764\" (UID: \"fb1dd6e9-8801-464e-952c-d345498b132a\") " pod="openstack/nova-cell1-conductor-db-sync-vd764" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.810582 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb1dd6e9-8801-464e-952c-d345498b132a-config-data\") pod \"nova-cell1-conductor-db-sync-vd764\" (UID: \"fb1dd6e9-8801-464e-952c-d345498b132a\") " pod="openstack/nova-cell1-conductor-db-sync-vd764" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.810703 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fmmn\" (UniqueName: \"kubernetes.io/projected/fb1dd6e9-8801-464e-952c-d345498b132a-kube-api-access-5fmmn\") pod \"nova-cell1-conductor-db-sync-vd764\" (UID: \"fb1dd6e9-8801-464e-952c-d345498b132a\") " pod="openstack/nova-cell1-conductor-db-sync-vd764" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.810918 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb1dd6e9-8801-464e-952c-d345498b132a-scripts\") pod \"nova-cell1-conductor-db-sync-vd764\" (UID: \"fb1dd6e9-8801-464e-952c-d345498b132a\") " pod="openstack/nova-cell1-conductor-db-sync-vd764" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.832742 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb1dd6e9-8801-464e-952c-d345498b132a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vd764\" (UID: \"fb1dd6e9-8801-464e-952c-d345498b132a\") " pod="openstack/nova-cell1-conductor-db-sync-vd764" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.833855 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.833924 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.862232 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fmmn\" (UniqueName: \"kubernetes.io/projected/fb1dd6e9-8801-464e-952c-d345498b132a-kube-api-access-5fmmn\") pod \"nova-cell1-conductor-db-sync-vd764\" (UID: \"fb1dd6e9-8801-464e-952c-d345498b132a\") " pod="openstack/nova-cell1-conductor-db-sync-vd764" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.862352 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb1dd6e9-8801-464e-952c-d345498b132a-config-data\") pod \"nova-cell1-conductor-db-sync-vd764\" (UID: \"fb1dd6e9-8801-464e-952c-d345498b132a\") " pod="openstack/nova-cell1-conductor-db-sync-vd764" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.864325 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb1dd6e9-8801-464e-952c-d345498b132a-scripts\") pod \"nova-cell1-conductor-db-sync-vd764\" (UID: \"fb1dd6e9-8801-464e-952c-d345498b132a\") " pod="openstack/nova-cell1-conductor-db-sync-vd764" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.899327 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.964919 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vd764" Mar 21 05:09:08 crc kubenswrapper[4775]: I0321 05:09:08.967414 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-qsbbk"] Mar 21 05:09:09 crc kubenswrapper[4775]: I0321 05:09:09.109429 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 21 05:09:09 crc kubenswrapper[4775]: I0321 05:09:09.491152 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vd764"] Mar 21 05:09:09 crc kubenswrapper[4775]: I0321 05:09:09.524359 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-pckp8" event={"ID":"5a407dd6-593f-4806-884b-8e031639d25d","Type":"ContainerStarted","Data":"650b9597595a69528b159834c598c6d1c2eb6858daf6ed558c1a6a03d2d450dd"} Mar 21 05:09:09 crc kubenswrapper[4775]: I0321 05:09:09.537242 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87dad904-a84b-4dcf-9caa-3fe89053b96b","Type":"ContainerStarted","Data":"3c21cc3a28ac294006c3d112d32cbbbabe9d8f787077b90e376062aeec6dab3e"} Mar 21 05:09:09 crc kubenswrapper[4775]: I0321 05:09:09.547978 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2","Type":"ContainerStarted","Data":"e7ffc178120764777a6e0c0217aa92bdaf948da3b3e7aaf4502f70ada80abfb2"} Mar 21 05:09:09 crc kubenswrapper[4775]: I0321 05:09:09.555099 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e114fa87-9e65-4a0b-bad1-81f521aeef85","Type":"ContainerStarted","Data":"8a454c45ca71508ee3e69e89e31c30561bb964893c09c7b3c20372f820d152ee"} Mar 21 05:09:09 crc kubenswrapper[4775]: I0321 05:09:09.557772 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d7219652-7280-4a81-b501-b6e41673162b","Type":"ContainerStarted","Data":"c67a6d4fa6caed2b2522033053bcb1027513121c3bd46425c786741079d6ba00"} Mar 21 05:09:09 crc kubenswrapper[4775]: I0321 05:09:09.566760 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-pckp8" podStartSLOduration=2.566735796 podStartE2EDuration="2.566735796s" podCreationTimestamp="2026-03-21 05:09:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:09:09.539718342 +0000 UTC m=+1302.516181966" watchObservedRunningTime="2026-03-21 05:09:09.566735796 +0000 UTC m=+1302.543199420" Mar 21 05:09:09 crc kubenswrapper[4775]: I0321 05:09:09.568184 4775 generic.go:334] "Generic (PLEG): container finished" podID="20a97ec2-24be-494f-a3e8-dc7d202021c3" containerID="992670c46bdc3ecfcf02d509f68efcc94de86d4d365af73eee5b378735e764a6" exitCode=0 Mar 21 05:09:09 crc kubenswrapper[4775]: I0321 05:09:09.568206 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" event={"ID":"20a97ec2-24be-494f-a3e8-dc7d202021c3","Type":"ContainerDied","Data":"992670c46bdc3ecfcf02d509f68efcc94de86d4d365af73eee5b378735e764a6"} Mar 21 05:09:09 crc kubenswrapper[4775]: I0321 05:09:09.569683 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" event={"ID":"20a97ec2-24be-494f-a3e8-dc7d202021c3","Type":"ContainerStarted","Data":"9a2e5054b120d9d9b114985bfe9a3a6397285644c20c24699a021ae58fcb7c61"} Mar 21 05:09:10 crc kubenswrapper[4775]: I0321 05:09:10.581141 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" event={"ID":"20a97ec2-24be-494f-a3e8-dc7d202021c3","Type":"ContainerStarted","Data":"210c52f72b30e6d06b1636b7f4ba7c385a206b219bca52c176d9da81b984a93a"} Mar 21 05:09:10 crc kubenswrapper[4775]: I0321 05:09:10.581483 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:10 crc kubenswrapper[4775]: I0321 05:09:10.587261 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vd764" event={"ID":"fb1dd6e9-8801-464e-952c-d345498b132a","Type":"ContainerStarted","Data":"166965bd5533ce39880e58f576322a3fbe5ebbddcee7067e02f0a62bbadf6013"} Mar 21 05:09:10 crc kubenswrapper[4775]: I0321 05:09:10.587308 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vd764" event={"ID":"fb1dd6e9-8801-464e-952c-d345498b132a","Type":"ContainerStarted","Data":"5e30087cecb9bbd6f3df2a364239f70f52ef2de20ec9c4dfd84fb90814b44306"} Mar 21 05:09:10 crc kubenswrapper[4775]: I0321 05:09:10.614759 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" podStartSLOduration=3.614736293 podStartE2EDuration="3.614736293s" podCreationTimestamp="2026-03-21 05:09:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:09:10.60507384 +0000 UTC m=+1303.581537474" watchObservedRunningTime="2026-03-21 05:09:10.614736293 +0000 UTC m=+1303.591199917" Mar 21 05:09:10 crc kubenswrapper[4775]: I0321 05:09:10.627065 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-vd764" podStartSLOduration=2.627041451 podStartE2EDuration="2.627041451s" podCreationTimestamp="2026-03-21 05:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:09:10.623430929 +0000 UTC m=+1303.599894563" watchObservedRunningTime="2026-03-21 05:09:10.627041451 +0000 UTC m=+1303.603505095" Mar 21 05:09:10 crc kubenswrapper[4775]: I0321 05:09:10.938859 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 21 05:09:10 crc kubenswrapper[4775]: I0321 05:09:10.948005 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:09:13 crc kubenswrapper[4775]: I0321 05:09:13.483291 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 21 05:09:13 crc kubenswrapper[4775]: I0321 05:09:13.620228 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2","Type":"ContainerStarted","Data":"bdf97a0cf8b488689d63ad976ec7a07bd6c1bed2c6f6a773ddf11243bbcc10aa"} Mar 21 05:09:13 crc kubenswrapper[4775]: I0321 05:09:13.620370 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://bdf97a0cf8b488689d63ad976ec7a07bd6c1bed2c6f6a773ddf11243bbcc10aa" gracePeriod=30 Mar 21 05:09:13 crc kubenswrapper[4775]: I0321 05:09:13.626737 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e114fa87-9e65-4a0b-bad1-81f521aeef85","Type":"ContainerStarted","Data":"4db8a2108514793dd399803fe0f901aa4854b0ae09ceaf27535063e2f431666a"} Mar 21 05:09:13 crc kubenswrapper[4775]: I0321 05:09:13.630045 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d7219652-7280-4a81-b501-b6e41673162b","Type":"ContainerStarted","Data":"2076143ea51df5acf1912050c6b923d311867381f6ad9a4e7c31e5ba00ded2cb"} Mar 21 05:09:13 crc kubenswrapper[4775]: I0321 05:09:13.630081 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d7219652-7280-4a81-b501-b6e41673162b","Type":"ContainerStarted","Data":"875af4c853efd2be80431c69406f831a833e919fb5a28d0763b186677c3a8433"} Mar 21 05:09:13 crc kubenswrapper[4775]: I0321 05:09:13.630213 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d7219652-7280-4a81-b501-b6e41673162b" containerName="nova-metadata-log" containerID="cri-o://875af4c853efd2be80431c69406f831a833e919fb5a28d0763b186677c3a8433" gracePeriod=30 Mar 21 05:09:13 crc kubenswrapper[4775]: I0321 05:09:13.630497 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d7219652-7280-4a81-b501-b6e41673162b" containerName="nova-metadata-metadata" containerID="cri-o://2076143ea51df5acf1912050c6b923d311867381f6ad9a4e7c31e5ba00ded2cb" gracePeriod=30 Mar 21 05:09:13 crc kubenswrapper[4775]: I0321 05:09:13.638720 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87dad904-a84b-4dcf-9caa-3fe89053b96b","Type":"ContainerStarted","Data":"3d41a9111ff02a22981deef2e158c7aaae04c761a3f477f9cde2ecff8054c0f1"} Mar 21 05:09:13 crc kubenswrapper[4775]: I0321 05:09:13.638771 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87dad904-a84b-4dcf-9caa-3fe89053b96b","Type":"ContainerStarted","Data":"a3e8af6598015c259cd8e37d94c33e2e14ca3a07bdbbfc459894647fdb9e40ea"} Mar 21 05:09:13 crc kubenswrapper[4775]: I0321 05:09:13.670734 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.061585605 podStartE2EDuration="6.670707805s" podCreationTimestamp="2026-03-21 05:09:07 +0000 UTC" firstStartedPulling="2026-03-21 05:09:09.119527053 +0000 UTC m=+1302.095990677" lastFinishedPulling="2026-03-21 05:09:12.728649253 +0000 UTC m=+1305.705112877" observedRunningTime="2026-03-21 05:09:13.668329987 +0000 UTC m=+1306.644793611" watchObservedRunningTime="2026-03-21 05:09:13.670707805 +0000 UTC m=+1306.647171429" Mar 21 05:09:13 crc kubenswrapper[4775]: I0321 05:09:13.692981 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.8407378420000002 podStartE2EDuration="6.692964484s" podCreationTimestamp="2026-03-21 05:09:07 +0000 UTC" firstStartedPulling="2026-03-21 05:09:08.882984236 +0000 UTC m=+1301.859447860" lastFinishedPulling="2026-03-21 05:09:12.735210878 +0000 UTC m=+1305.711674502" observedRunningTime="2026-03-21 05:09:13.690022971 +0000 UTC m=+1306.666486595" watchObservedRunningTime="2026-03-21 05:09:13.692964484 +0000 UTC m=+1306.669428108" Mar 21 05:09:13 crc kubenswrapper[4775]: I0321 05:09:13.712175 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.713941068 podStartE2EDuration="6.712159547s" podCreationTimestamp="2026-03-21 05:09:07 +0000 UTC" firstStartedPulling="2026-03-21 05:09:08.730365342 +0000 UTC m=+1301.706828966" lastFinishedPulling="2026-03-21 05:09:12.728583821 +0000 UTC m=+1305.705047445" observedRunningTime="2026-03-21 05:09:13.710271423 +0000 UTC m=+1306.686735047" watchObservedRunningTime="2026-03-21 05:09:13.712159547 +0000 UTC m=+1306.688623171" Mar 21 05:09:13 crc kubenswrapper[4775]: I0321 05:09:13.734971 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.8688467060000002 podStartE2EDuration="6.734952871s" podCreationTimestamp="2026-03-21 05:09:07 +0000 UTC" firstStartedPulling="2026-03-21 05:09:08.862478946 +0000 UTC m=+1301.838942570" lastFinishedPulling="2026-03-21 05:09:12.728585111 +0000 UTC m=+1305.705048735" observedRunningTime="2026-03-21 05:09:13.734633502 +0000 UTC m=+1306.711097126" watchObservedRunningTime="2026-03-21 05:09:13.734952871 +0000 UTC m=+1306.711416495" Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.609754 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.658476 4775 generic.go:334] "Generic (PLEG): container finished" podID="d7219652-7280-4a81-b501-b6e41673162b" containerID="2076143ea51df5acf1912050c6b923d311867381f6ad9a4e7c31e5ba00ded2cb" exitCode=0 Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.658814 4775 generic.go:334] "Generic (PLEG): container finished" podID="d7219652-7280-4a81-b501-b6e41673162b" containerID="875af4c853efd2be80431c69406f831a833e919fb5a28d0763b186677c3a8433" exitCode=143 Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.658565 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.658578 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d7219652-7280-4a81-b501-b6e41673162b","Type":"ContainerDied","Data":"2076143ea51df5acf1912050c6b923d311867381f6ad9a4e7c31e5ba00ded2cb"} Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.658921 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d7219652-7280-4a81-b501-b6e41673162b","Type":"ContainerDied","Data":"875af4c853efd2be80431c69406f831a833e919fb5a28d0763b186677c3a8433"} Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.658935 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d7219652-7280-4a81-b501-b6e41673162b","Type":"ContainerDied","Data":"c67a6d4fa6caed2b2522033053bcb1027513121c3bd46425c786741079d6ba00"} Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.658952 4775 scope.go:117] "RemoveContainer" containerID="2076143ea51df5acf1912050c6b923d311867381f6ad9a4e7c31e5ba00ded2cb" Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.664789 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7219652-7280-4a81-b501-b6e41673162b-logs\") pod \"d7219652-7280-4a81-b501-b6e41673162b\" (UID: \"d7219652-7280-4a81-b501-b6e41673162b\") " Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.664899 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9m8c\" (UniqueName: \"kubernetes.io/projected/d7219652-7280-4a81-b501-b6e41673162b-kube-api-access-g9m8c\") pod \"d7219652-7280-4a81-b501-b6e41673162b\" (UID: \"d7219652-7280-4a81-b501-b6e41673162b\") " Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.665038 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7219652-7280-4a81-b501-b6e41673162b-combined-ca-bundle\") pod \"d7219652-7280-4a81-b501-b6e41673162b\" (UID: \"d7219652-7280-4a81-b501-b6e41673162b\") " Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.665064 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7219652-7280-4a81-b501-b6e41673162b-config-data\") pod \"d7219652-7280-4a81-b501-b6e41673162b\" (UID: \"d7219652-7280-4a81-b501-b6e41673162b\") " Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.665453 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7219652-7280-4a81-b501-b6e41673162b-logs" (OuterVolumeSpecName: "logs") pod "d7219652-7280-4a81-b501-b6e41673162b" (UID: "d7219652-7280-4a81-b501-b6e41673162b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.666245 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7219652-7280-4a81-b501-b6e41673162b-logs\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.674598 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7219652-7280-4a81-b501-b6e41673162b-kube-api-access-g9m8c" (OuterVolumeSpecName: "kube-api-access-g9m8c") pod "d7219652-7280-4a81-b501-b6e41673162b" (UID: "d7219652-7280-4a81-b501-b6e41673162b"). InnerVolumeSpecName "kube-api-access-g9m8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.687862 4775 scope.go:117] "RemoveContainer" containerID="875af4c853efd2be80431c69406f831a833e919fb5a28d0763b186677c3a8433" Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.698307 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7219652-7280-4a81-b501-b6e41673162b-config-data" (OuterVolumeSpecName: "config-data") pod "d7219652-7280-4a81-b501-b6e41673162b" (UID: "d7219652-7280-4a81-b501-b6e41673162b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.706291 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7219652-7280-4a81-b501-b6e41673162b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7219652-7280-4a81-b501-b6e41673162b" (UID: "d7219652-7280-4a81-b501-b6e41673162b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.715790 4775 scope.go:117] "RemoveContainer" containerID="2076143ea51df5acf1912050c6b923d311867381f6ad9a4e7c31e5ba00ded2cb" Mar 21 05:09:14 crc kubenswrapper[4775]: E0321 05:09:14.716262 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2076143ea51df5acf1912050c6b923d311867381f6ad9a4e7c31e5ba00ded2cb\": container with ID starting with 2076143ea51df5acf1912050c6b923d311867381f6ad9a4e7c31e5ba00ded2cb not found: ID does not exist" containerID="2076143ea51df5acf1912050c6b923d311867381f6ad9a4e7c31e5ba00ded2cb" Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.716320 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2076143ea51df5acf1912050c6b923d311867381f6ad9a4e7c31e5ba00ded2cb"} err="failed to get container status \"2076143ea51df5acf1912050c6b923d311867381f6ad9a4e7c31e5ba00ded2cb\": rpc error: code = NotFound desc = could not find container \"2076143ea51df5acf1912050c6b923d311867381f6ad9a4e7c31e5ba00ded2cb\": container with ID starting with 2076143ea51df5acf1912050c6b923d311867381f6ad9a4e7c31e5ba00ded2cb not found: ID does not exist" Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.716359 4775 scope.go:117] "RemoveContainer" containerID="875af4c853efd2be80431c69406f831a833e919fb5a28d0763b186677c3a8433" Mar 21 05:09:14 crc kubenswrapper[4775]: E0321 05:09:14.716760 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"875af4c853efd2be80431c69406f831a833e919fb5a28d0763b186677c3a8433\": container with ID starting with 875af4c853efd2be80431c69406f831a833e919fb5a28d0763b186677c3a8433 not found: ID does not exist" containerID="875af4c853efd2be80431c69406f831a833e919fb5a28d0763b186677c3a8433" Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.716817 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"875af4c853efd2be80431c69406f831a833e919fb5a28d0763b186677c3a8433"} err="failed to get container status \"875af4c853efd2be80431c69406f831a833e919fb5a28d0763b186677c3a8433\": rpc error: code = NotFound desc = could not find container \"875af4c853efd2be80431c69406f831a833e919fb5a28d0763b186677c3a8433\": container with ID starting with 875af4c853efd2be80431c69406f831a833e919fb5a28d0763b186677c3a8433 not found: ID does not exist" Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.716848 4775 scope.go:117] "RemoveContainer" containerID="2076143ea51df5acf1912050c6b923d311867381f6ad9a4e7c31e5ba00ded2cb" Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.717378 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2076143ea51df5acf1912050c6b923d311867381f6ad9a4e7c31e5ba00ded2cb"} err="failed to get container status \"2076143ea51df5acf1912050c6b923d311867381f6ad9a4e7c31e5ba00ded2cb\": rpc error: code = NotFound desc = could not find container \"2076143ea51df5acf1912050c6b923d311867381f6ad9a4e7c31e5ba00ded2cb\": container with ID starting with 2076143ea51df5acf1912050c6b923d311867381f6ad9a4e7c31e5ba00ded2cb not found: ID does not exist" Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.717408 4775 scope.go:117] "RemoveContainer" containerID="875af4c853efd2be80431c69406f831a833e919fb5a28d0763b186677c3a8433" Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.717854 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"875af4c853efd2be80431c69406f831a833e919fb5a28d0763b186677c3a8433"} err="failed to get container status \"875af4c853efd2be80431c69406f831a833e919fb5a28d0763b186677c3a8433\": rpc error: code = NotFound desc = could not find container \"875af4c853efd2be80431c69406f831a833e919fb5a28d0763b186677c3a8433\": container with ID starting with 875af4c853efd2be80431c69406f831a833e919fb5a28d0763b186677c3a8433 not found: ID does not exist" Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.768224 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9m8c\" (UniqueName: \"kubernetes.io/projected/d7219652-7280-4a81-b501-b6e41673162b-kube-api-access-g9m8c\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.768255 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7219652-7280-4a81-b501-b6e41673162b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:14 crc kubenswrapper[4775]: I0321 05:09:14.768266 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7219652-7280-4a81-b501-b6e41673162b-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.011443 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.024189 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.036181 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:09:15 crc kubenswrapper[4775]: E0321 05:09:15.036637 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7219652-7280-4a81-b501-b6e41673162b" containerName="nova-metadata-metadata" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.036663 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7219652-7280-4a81-b501-b6e41673162b" containerName="nova-metadata-metadata" Mar 21 05:09:15 crc kubenswrapper[4775]: E0321 05:09:15.036768 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7219652-7280-4a81-b501-b6e41673162b" containerName="nova-metadata-log" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.036779 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7219652-7280-4a81-b501-b6e41673162b" containerName="nova-metadata-log" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.036973 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7219652-7280-4a81-b501-b6e41673162b" containerName="nova-metadata-metadata" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.037000 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7219652-7280-4a81-b501-b6e41673162b" containerName="nova-metadata-log" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.038015 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.040805 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.042443 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.067907 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.073381 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82600d19-b64e-4737-8a1f-ff1289daa3ed-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"82600d19-b64e-4737-8a1f-ff1289daa3ed\") " pod="openstack/nova-metadata-0" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.073643 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82600d19-b64e-4737-8a1f-ff1289daa3ed-logs\") pod \"nova-metadata-0\" (UID: \"82600d19-b64e-4737-8a1f-ff1289daa3ed\") " pod="openstack/nova-metadata-0" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.073823 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82600d19-b64e-4737-8a1f-ff1289daa3ed-config-data\") pod \"nova-metadata-0\" (UID: \"82600d19-b64e-4737-8a1f-ff1289daa3ed\") " pod="openstack/nova-metadata-0" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.073882 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsw75\" (UniqueName: \"kubernetes.io/projected/82600d19-b64e-4737-8a1f-ff1289daa3ed-kube-api-access-jsw75\") pod \"nova-metadata-0\" (UID: \"82600d19-b64e-4737-8a1f-ff1289daa3ed\") " pod="openstack/nova-metadata-0" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.073906 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82600d19-b64e-4737-8a1f-ff1289daa3ed-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"82600d19-b64e-4737-8a1f-ff1289daa3ed\") " pod="openstack/nova-metadata-0" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.180111 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82600d19-b64e-4737-8a1f-ff1289daa3ed-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"82600d19-b64e-4737-8a1f-ff1289daa3ed\") " pod="openstack/nova-metadata-0" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.180199 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82600d19-b64e-4737-8a1f-ff1289daa3ed-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"82600d19-b64e-4737-8a1f-ff1289daa3ed\") " pod="openstack/nova-metadata-0" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.180274 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82600d19-b64e-4737-8a1f-ff1289daa3ed-logs\") pod \"nova-metadata-0\" (UID: \"82600d19-b64e-4737-8a1f-ff1289daa3ed\") " pod="openstack/nova-metadata-0" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.180332 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82600d19-b64e-4737-8a1f-ff1289daa3ed-config-data\") pod \"nova-metadata-0\" (UID: \"82600d19-b64e-4737-8a1f-ff1289daa3ed\") " pod="openstack/nova-metadata-0" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.180362 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsw75\" (UniqueName: \"kubernetes.io/projected/82600d19-b64e-4737-8a1f-ff1289daa3ed-kube-api-access-jsw75\") pod \"nova-metadata-0\" (UID: \"82600d19-b64e-4737-8a1f-ff1289daa3ed\") " pod="openstack/nova-metadata-0" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.180872 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82600d19-b64e-4737-8a1f-ff1289daa3ed-logs\") pod \"nova-metadata-0\" (UID: \"82600d19-b64e-4737-8a1f-ff1289daa3ed\") " pod="openstack/nova-metadata-0" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.184052 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82600d19-b64e-4737-8a1f-ff1289daa3ed-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"82600d19-b64e-4737-8a1f-ff1289daa3ed\") " pod="openstack/nova-metadata-0" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.184192 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82600d19-b64e-4737-8a1f-ff1289daa3ed-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"82600d19-b64e-4737-8a1f-ff1289daa3ed\") " pod="openstack/nova-metadata-0" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.198834 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsw75\" (UniqueName: \"kubernetes.io/projected/82600d19-b64e-4737-8a1f-ff1289daa3ed-kube-api-access-jsw75\") pod \"nova-metadata-0\" (UID: \"82600d19-b64e-4737-8a1f-ff1289daa3ed\") " pod="openstack/nova-metadata-0" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.198942 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82600d19-b64e-4737-8a1f-ff1289daa3ed-config-data\") pod \"nova-metadata-0\" (UID: \"82600d19-b64e-4737-8a1f-ff1289daa3ed\") " pod="openstack/nova-metadata-0" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.356562 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.677526 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7219652-7280-4a81-b501-b6e41673162b" path="/var/lib/kubelet/pods/d7219652-7280-4a81-b501-b6e41673162b/volumes" Mar 21 05:09:15 crc kubenswrapper[4775]: I0321 05:09:15.830842 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:09:16 crc kubenswrapper[4775]: I0321 05:09:16.690205 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82600d19-b64e-4737-8a1f-ff1289daa3ed","Type":"ContainerStarted","Data":"a9e328b0de38c9a70bb3528a6b0dd0685f8454aa009ccc97d323977d2e868a06"} Mar 21 05:09:16 crc kubenswrapper[4775]: I0321 05:09:16.690577 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82600d19-b64e-4737-8a1f-ff1289daa3ed","Type":"ContainerStarted","Data":"2cc3d35e715a6db1c79daff232799e16bf15b4a6f8ae63e1656332d4536a706b"} Mar 21 05:09:16 crc kubenswrapper[4775]: I0321 05:09:16.690599 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82600d19-b64e-4737-8a1f-ff1289daa3ed","Type":"ContainerStarted","Data":"1aef16226ce61b14de14608541f12818545fa3117a238adb67e56a57bcd74341"} Mar 21 05:09:16 crc kubenswrapper[4775]: I0321 05:09:16.741525 4775 generic.go:334] "Generic (PLEG): container finished" podID="5a407dd6-593f-4806-884b-8e031639d25d" containerID="650b9597595a69528b159834c598c6d1c2eb6858daf6ed558c1a6a03d2d450dd" exitCode=0 Mar 21 05:09:16 crc kubenswrapper[4775]: I0321 05:09:16.741591 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-pckp8" event={"ID":"5a407dd6-593f-4806-884b-8e031639d25d","Type":"ContainerDied","Data":"650b9597595a69528b159834c598c6d1c2eb6858daf6ed558c1a6a03d2d450dd"} Mar 21 05:09:16 crc kubenswrapper[4775]: I0321 05:09:16.797687 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.797665573 podStartE2EDuration="1.797665573s" podCreationTimestamp="2026-03-21 05:09:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:09:16.762161729 +0000 UTC m=+1309.738625373" watchObservedRunningTime="2026-03-21 05:09:16.797665573 +0000 UTC m=+1309.774129197" Mar 21 05:09:17 crc kubenswrapper[4775]: I0321 05:09:17.937427 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 21 05:09:17 crc kubenswrapper[4775]: I0321 05:09:17.937872 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.055245 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.055313 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.086823 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.104369 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.115137 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-pckp8" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.138468 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a407dd6-593f-4806-884b-8e031639d25d-scripts\") pod \"5a407dd6-593f-4806-884b-8e031639d25d\" (UID: \"5a407dd6-593f-4806-884b-8e031639d25d\") " Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.138597 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a407dd6-593f-4806-884b-8e031639d25d-config-data\") pod \"5a407dd6-593f-4806-884b-8e031639d25d\" (UID: \"5a407dd6-593f-4806-884b-8e031639d25d\") " Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.138647 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbdhj\" (UniqueName: \"kubernetes.io/projected/5a407dd6-593f-4806-884b-8e031639d25d-kube-api-access-zbdhj\") pod \"5a407dd6-593f-4806-884b-8e031639d25d\" (UID: \"5a407dd6-593f-4806-884b-8e031639d25d\") " Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.138722 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a407dd6-593f-4806-884b-8e031639d25d-combined-ca-bundle\") pod \"5a407dd6-593f-4806-884b-8e031639d25d\" (UID: \"5a407dd6-593f-4806-884b-8e031639d25d\") " Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.160277 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a407dd6-593f-4806-884b-8e031639d25d-kube-api-access-zbdhj" (OuterVolumeSpecName: "kube-api-access-zbdhj") pod "5a407dd6-593f-4806-884b-8e031639d25d" (UID: "5a407dd6-593f-4806-884b-8e031639d25d"). InnerVolumeSpecName "kube-api-access-zbdhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.190358 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a407dd6-593f-4806-884b-8e031639d25d-scripts" (OuterVolumeSpecName: "scripts") pod "5a407dd6-593f-4806-884b-8e031639d25d" (UID: "5a407dd6-593f-4806-884b-8e031639d25d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.205579 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a407dd6-593f-4806-884b-8e031639d25d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a407dd6-593f-4806-884b-8e031639d25d" (UID: "5a407dd6-593f-4806-884b-8e031639d25d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.224239 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a407dd6-593f-4806-884b-8e031639d25d-config-data" (OuterVolumeSpecName: "config-data") pod "5a407dd6-593f-4806-884b-8e031639d25d" (UID: "5a407dd6-593f-4806-884b-8e031639d25d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.224606 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-w2ltw"] Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.225080 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" podUID="d7dc929b-501f-43cb-8811-394841cc54f9" containerName="dnsmasq-dns" containerID="cri-o://07d83e49b0e806d759a2ee326367b45aab2ed5d87b3ee996d3b75ec5f20c736e" gracePeriod=10 Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.242639 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a407dd6-593f-4806-884b-8e031639d25d-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.242699 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbdhj\" (UniqueName: \"kubernetes.io/projected/5a407dd6-593f-4806-884b-8e031639d25d-kube-api-access-zbdhj\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.242715 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a407dd6-593f-4806-884b-8e031639d25d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.242725 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a407dd6-593f-4806-884b-8e031639d25d-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.246107 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.688672 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.752908 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-ovsdbserver-sb\") pod \"d7dc929b-501f-43cb-8811-394841cc54f9\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.753171 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbc62\" (UniqueName: \"kubernetes.io/projected/d7dc929b-501f-43cb-8811-394841cc54f9-kube-api-access-tbc62\") pod \"d7dc929b-501f-43cb-8811-394841cc54f9\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.753417 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-dns-svc\") pod \"d7dc929b-501f-43cb-8811-394841cc54f9\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.753450 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-dns-swift-storage-0\") pod \"d7dc929b-501f-43cb-8811-394841cc54f9\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.753650 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-ovsdbserver-nb\") pod \"d7dc929b-501f-43cb-8811-394841cc54f9\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.753802 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-config\") pod \"d7dc929b-501f-43cb-8811-394841cc54f9\" (UID: \"d7dc929b-501f-43cb-8811-394841cc54f9\") " Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.766942 4775 generic.go:334] "Generic (PLEG): container finished" podID="d7dc929b-501f-43cb-8811-394841cc54f9" containerID="07d83e49b0e806d759a2ee326367b45aab2ed5d87b3ee996d3b75ec5f20c736e" exitCode=0 Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.767060 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" event={"ID":"d7dc929b-501f-43cb-8811-394841cc54f9","Type":"ContainerDied","Data":"07d83e49b0e806d759a2ee326367b45aab2ed5d87b3ee996d3b75ec5f20c736e"} Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.767102 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" event={"ID":"d7dc929b-501f-43cb-8811-394841cc54f9","Type":"ContainerDied","Data":"4431b638594d2b34fefc401fb2d4af3e48aa80179d99ea9d13c2365707189fa3"} Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.767144 4775 scope.go:117] "RemoveContainer" containerID="07d83e49b0e806d759a2ee326367b45aab2ed5d87b3ee996d3b75ec5f20c736e" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.767630 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-w2ltw" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.776825 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-pckp8" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.776966 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7dc929b-501f-43cb-8811-394841cc54f9-kube-api-access-tbc62" (OuterVolumeSpecName: "kube-api-access-tbc62") pod "d7dc929b-501f-43cb-8811-394841cc54f9" (UID: "d7dc929b-501f-43cb-8811-394841cc54f9"). InnerVolumeSpecName "kube-api-access-tbc62". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.777027 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-pckp8" event={"ID":"5a407dd6-593f-4806-884b-8e031639d25d","Type":"ContainerDied","Data":"040855e862320088965a0b53c32aefe654235d67e7c2aec1b7da6cf02854cd90"} Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.777070 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="040855e862320088965a0b53c32aefe654235d67e7c2aec1b7da6cf02854cd90" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.797581 4775 scope.go:117] "RemoveContainer" containerID="6ab789fb8fabdc5b2c75b49b35b369108626a6a98dbf396798f603e25b157d25" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.826489 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-config" (OuterVolumeSpecName: "config") pod "d7dc929b-501f-43cb-8811-394841cc54f9" (UID: "d7dc929b-501f-43cb-8811-394841cc54f9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.847543 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d7dc929b-501f-43cb-8811-394841cc54f9" (UID: "d7dc929b-501f-43cb-8811-394841cc54f9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.850829 4775 scope.go:117] "RemoveContainer" containerID="07d83e49b0e806d759a2ee326367b45aab2ed5d87b3ee996d3b75ec5f20c736e" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.851938 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 21 05:09:18 crc kubenswrapper[4775]: E0321 05:09:18.859907 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07d83e49b0e806d759a2ee326367b45aab2ed5d87b3ee996d3b75ec5f20c736e\": container with ID starting with 07d83e49b0e806d759a2ee326367b45aab2ed5d87b3ee996d3b75ec5f20c736e not found: ID does not exist" containerID="07d83e49b0e806d759a2ee326367b45aab2ed5d87b3ee996d3b75ec5f20c736e" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.859951 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07d83e49b0e806d759a2ee326367b45aab2ed5d87b3ee996d3b75ec5f20c736e"} err="failed to get container status \"07d83e49b0e806d759a2ee326367b45aab2ed5d87b3ee996d3b75ec5f20c736e\": rpc error: code = NotFound desc = could not find container \"07d83e49b0e806d759a2ee326367b45aab2ed5d87b3ee996d3b75ec5f20c736e\": container with ID starting with 07d83e49b0e806d759a2ee326367b45aab2ed5d87b3ee996d3b75ec5f20c736e not found: ID does not exist" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.859974 4775 scope.go:117] "RemoveContainer" containerID="6ab789fb8fabdc5b2c75b49b35b369108626a6a98dbf396798f603e25b157d25" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.860318 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbc62\" (UniqueName: \"kubernetes.io/projected/d7dc929b-501f-43cb-8811-394841cc54f9-kube-api-access-tbc62\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.860364 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.860379 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:18 crc kubenswrapper[4775]: E0321 05:09:18.860533 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ab789fb8fabdc5b2c75b49b35b369108626a6a98dbf396798f603e25b157d25\": container with ID starting with 6ab789fb8fabdc5b2c75b49b35b369108626a6a98dbf396798f603e25b157d25 not found: ID does not exist" containerID="6ab789fb8fabdc5b2c75b49b35b369108626a6a98dbf396798f603e25b157d25" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.860556 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ab789fb8fabdc5b2c75b49b35b369108626a6a98dbf396798f603e25b157d25"} err="failed to get container status \"6ab789fb8fabdc5b2c75b49b35b369108626a6a98dbf396798f603e25b157d25\": rpc error: code = NotFound desc = could not find container \"6ab789fb8fabdc5b2c75b49b35b369108626a6a98dbf396798f603e25b157d25\": container with ID starting with 6ab789fb8fabdc5b2c75b49b35b369108626a6a98dbf396798f603e25b157d25 not found: ID does not exist" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.860580 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d7dc929b-501f-43cb-8811-394841cc54f9" (UID: "d7dc929b-501f-43cb-8811-394841cc54f9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.863355 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d7dc929b-501f-43cb-8811-394841cc54f9" (UID: "d7dc929b-501f-43cb-8811-394841cc54f9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.865263 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d7dc929b-501f-43cb-8811-394841cc54f9" (UID: "d7dc929b-501f-43cb-8811-394841cc54f9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.938745 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.952531 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.952790 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="87dad904-a84b-4dcf-9caa-3fe89053b96b" containerName="nova-api-log" containerID="cri-o://a3e8af6598015c259cd8e37d94c33e2e14ca3a07bdbbfc459894647fdb9e40ea" gracePeriod=30 Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.952938 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="87dad904-a84b-4dcf-9caa-3fe89053b96b" containerName="nova-api-api" containerID="cri-o://3d41a9111ff02a22981deef2e158c7aaae04c761a3f477f9cde2ecff8054c0f1" gracePeriod=30 Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.961903 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.961941 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.961952 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7dc929b-501f-43cb-8811-394841cc54f9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.963722 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="87dad904-a84b-4dcf-9caa-3fe89053b96b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.187:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.963888 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.964041 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="87dad904-a84b-4dcf-9caa-3fe89053b96b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.187:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.964114 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="82600d19-b64e-4737-8a1f-ff1289daa3ed" containerName="nova-metadata-log" containerID="cri-o://2cc3d35e715a6db1c79daff232799e16bf15b4a6f8ae63e1656332d4536a706b" gracePeriod=30 Mar 21 05:09:18 crc kubenswrapper[4775]: I0321 05:09:18.964318 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="82600d19-b64e-4737-8a1f-ff1289daa3ed" containerName="nova-metadata-metadata" containerID="cri-o://a9e328b0de38c9a70bb3528a6b0dd0685f8454aa009ccc97d323977d2e868a06" gracePeriod=30 Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.131194 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-w2ltw"] Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.138536 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-w2ltw"] Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.685879 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7dc929b-501f-43cb-8811-394841cc54f9" path="/var/lib/kubelet/pods/d7dc929b-501f-43cb-8811-394841cc54f9/volumes" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.739430 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.794397 4775 generic.go:334] "Generic (PLEG): container finished" podID="87dad904-a84b-4dcf-9caa-3fe89053b96b" containerID="a3e8af6598015c259cd8e37d94c33e2e14ca3a07bdbbfc459894647fdb9e40ea" exitCode=143 Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.794460 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87dad904-a84b-4dcf-9caa-3fe89053b96b","Type":"ContainerDied","Data":"a3e8af6598015c259cd8e37d94c33e2e14ca3a07bdbbfc459894647fdb9e40ea"} Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.797499 4775 generic.go:334] "Generic (PLEG): container finished" podID="82600d19-b64e-4737-8a1f-ff1289daa3ed" containerID="a9e328b0de38c9a70bb3528a6b0dd0685f8454aa009ccc97d323977d2e868a06" exitCode=0 Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.797518 4775 generic.go:334] "Generic (PLEG): container finished" podID="82600d19-b64e-4737-8a1f-ff1289daa3ed" containerID="2cc3d35e715a6db1c79daff232799e16bf15b4a6f8ae63e1656332d4536a706b" exitCode=143 Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.797555 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82600d19-b64e-4737-8a1f-ff1289daa3ed","Type":"ContainerDied","Data":"a9e328b0de38c9a70bb3528a6b0dd0685f8454aa009ccc97d323977d2e868a06"} Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.797577 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82600d19-b64e-4737-8a1f-ff1289daa3ed","Type":"ContainerDied","Data":"2cc3d35e715a6db1c79daff232799e16bf15b4a6f8ae63e1656332d4536a706b"} Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.797592 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82600d19-b64e-4737-8a1f-ff1289daa3ed","Type":"ContainerDied","Data":"1aef16226ce61b14de14608541f12818545fa3117a238adb67e56a57bcd74341"} Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.797614 4775 scope.go:117] "RemoveContainer" containerID="a9e328b0de38c9a70bb3528a6b0dd0685f8454aa009ccc97d323977d2e868a06" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.797923 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.798833 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.803684 4775 generic.go:334] "Generic (PLEG): container finished" podID="fb1dd6e9-8801-464e-952c-d345498b132a" containerID="166965bd5533ce39880e58f576322a3fbe5ebbddcee7067e02f0a62bbadf6013" exitCode=0 Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.803738 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vd764" event={"ID":"fb1dd6e9-8801-464e-952c-d345498b132a","Type":"ContainerDied","Data":"166965bd5533ce39880e58f576322a3fbe5ebbddcee7067e02f0a62bbadf6013"} Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.809241 4775 generic.go:334] "Generic (PLEG): container finished" podID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerID="30455e621655fa15c45c8fad2a223bc59355675a9f5b3da03f27d5155ab96f2d" exitCode=137 Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.810011 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.810212 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab91b8d2-32c9-4542-a730-7fc84e9892bd","Type":"ContainerDied","Data":"30455e621655fa15c45c8fad2a223bc59355675a9f5b3da03f27d5155ab96f2d"} Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.837696 4775 scope.go:117] "RemoveContainer" containerID="2cc3d35e715a6db1c79daff232799e16bf15b4a6f8ae63e1656332d4536a706b" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.866852 4775 scope.go:117] "RemoveContainer" containerID="a9e328b0de38c9a70bb3528a6b0dd0685f8454aa009ccc97d323977d2e868a06" Mar 21 05:09:19 crc kubenswrapper[4775]: E0321 05:09:19.869592 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9e328b0de38c9a70bb3528a6b0dd0685f8454aa009ccc97d323977d2e868a06\": container with ID starting with a9e328b0de38c9a70bb3528a6b0dd0685f8454aa009ccc97d323977d2e868a06 not found: ID does not exist" containerID="a9e328b0de38c9a70bb3528a6b0dd0685f8454aa009ccc97d323977d2e868a06" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.869728 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9e328b0de38c9a70bb3528a6b0dd0685f8454aa009ccc97d323977d2e868a06"} err="failed to get container status \"a9e328b0de38c9a70bb3528a6b0dd0685f8454aa009ccc97d323977d2e868a06\": rpc error: code = NotFound desc = could not find container \"a9e328b0de38c9a70bb3528a6b0dd0685f8454aa009ccc97d323977d2e868a06\": container with ID starting with a9e328b0de38c9a70bb3528a6b0dd0685f8454aa009ccc97d323977d2e868a06 not found: ID does not exist" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.869834 4775 scope.go:117] "RemoveContainer" containerID="2cc3d35e715a6db1c79daff232799e16bf15b4a6f8ae63e1656332d4536a706b" Mar 21 05:09:19 crc kubenswrapper[4775]: E0321 05:09:19.870397 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cc3d35e715a6db1c79daff232799e16bf15b4a6f8ae63e1656332d4536a706b\": container with ID starting with 2cc3d35e715a6db1c79daff232799e16bf15b4a6f8ae63e1656332d4536a706b not found: ID does not exist" containerID="2cc3d35e715a6db1c79daff232799e16bf15b4a6f8ae63e1656332d4536a706b" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.870444 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cc3d35e715a6db1c79daff232799e16bf15b4a6f8ae63e1656332d4536a706b"} err="failed to get container status \"2cc3d35e715a6db1c79daff232799e16bf15b4a6f8ae63e1656332d4536a706b\": rpc error: code = NotFound desc = could not find container \"2cc3d35e715a6db1c79daff232799e16bf15b4a6f8ae63e1656332d4536a706b\": container with ID starting with 2cc3d35e715a6db1c79daff232799e16bf15b4a6f8ae63e1656332d4536a706b not found: ID does not exist" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.870467 4775 scope.go:117] "RemoveContainer" containerID="a9e328b0de38c9a70bb3528a6b0dd0685f8454aa009ccc97d323977d2e868a06" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.871425 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9e328b0de38c9a70bb3528a6b0dd0685f8454aa009ccc97d323977d2e868a06"} err="failed to get container status \"a9e328b0de38c9a70bb3528a6b0dd0685f8454aa009ccc97d323977d2e868a06\": rpc error: code = NotFound desc = could not find container \"a9e328b0de38c9a70bb3528a6b0dd0685f8454aa009ccc97d323977d2e868a06\": container with ID starting with a9e328b0de38c9a70bb3528a6b0dd0685f8454aa009ccc97d323977d2e868a06 not found: ID does not exist" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.871458 4775 scope.go:117] "RemoveContainer" containerID="2cc3d35e715a6db1c79daff232799e16bf15b4a6f8ae63e1656332d4536a706b" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.872777 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cc3d35e715a6db1c79daff232799e16bf15b4a6f8ae63e1656332d4536a706b"} err="failed to get container status \"2cc3d35e715a6db1c79daff232799e16bf15b4a6f8ae63e1656332d4536a706b\": rpc error: code = NotFound desc = could not find container \"2cc3d35e715a6db1c79daff232799e16bf15b4a6f8ae63e1656332d4536a706b\": container with ID starting with 2cc3d35e715a6db1c79daff232799e16bf15b4a6f8ae63e1656332d4536a706b not found: ID does not exist" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.872802 4775 scope.go:117] "RemoveContainer" containerID="30455e621655fa15c45c8fad2a223bc59355675a9f5b3da03f27d5155ab96f2d" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.893849 4775 scope.go:117] "RemoveContainer" containerID="756b7ac25394ff7e03c954f4dded0675f6295888345b0b7a098c3ad32426a8ef" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.904190 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsw75\" (UniqueName: \"kubernetes.io/projected/82600d19-b64e-4737-8a1f-ff1289daa3ed-kube-api-access-jsw75\") pod \"82600d19-b64e-4737-8a1f-ff1289daa3ed\" (UID: \"82600d19-b64e-4737-8a1f-ff1289daa3ed\") " Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.905235 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-combined-ca-bundle\") pod \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.905334 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-config-data\") pod \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.905902 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-scripts\") pod \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.906015 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-sg-core-conf-yaml\") pod \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.906063 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82600d19-b64e-4737-8a1f-ff1289daa3ed-config-data\") pod \"82600d19-b64e-4737-8a1f-ff1289daa3ed\" (UID: \"82600d19-b64e-4737-8a1f-ff1289daa3ed\") " Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.906089 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82600d19-b64e-4737-8a1f-ff1289daa3ed-nova-metadata-tls-certs\") pod \"82600d19-b64e-4737-8a1f-ff1289daa3ed\" (UID: \"82600d19-b64e-4737-8a1f-ff1289daa3ed\") " Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.906145 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82600d19-b64e-4737-8a1f-ff1289daa3ed-logs\") pod \"82600d19-b64e-4737-8a1f-ff1289daa3ed\" (UID: \"82600d19-b64e-4737-8a1f-ff1289daa3ed\") " Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.906190 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82600d19-b64e-4737-8a1f-ff1289daa3ed-combined-ca-bundle\") pod \"82600d19-b64e-4737-8a1f-ff1289daa3ed\" (UID: \"82600d19-b64e-4737-8a1f-ff1289daa3ed\") " Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.906214 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab91b8d2-32c9-4542-a730-7fc84e9892bd-log-httpd\") pod \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.906234 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab91b8d2-32c9-4542-a730-7fc84e9892bd-run-httpd\") pod \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.906269 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4msqv\" (UniqueName: \"kubernetes.io/projected/ab91b8d2-32c9-4542-a730-7fc84e9892bd-kube-api-access-4msqv\") pod \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\" (UID: \"ab91b8d2-32c9-4542-a730-7fc84e9892bd\") " Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.907034 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82600d19-b64e-4737-8a1f-ff1289daa3ed-logs" (OuterVolumeSpecName: "logs") pod "82600d19-b64e-4737-8a1f-ff1289daa3ed" (UID: "82600d19-b64e-4737-8a1f-ff1289daa3ed"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.907410 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82600d19-b64e-4737-8a1f-ff1289daa3ed-logs\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.907681 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab91b8d2-32c9-4542-a730-7fc84e9892bd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ab91b8d2-32c9-4542-a730-7fc84e9892bd" (UID: "ab91b8d2-32c9-4542-a730-7fc84e9892bd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.907739 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab91b8d2-32c9-4542-a730-7fc84e9892bd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ab91b8d2-32c9-4542-a730-7fc84e9892bd" (UID: "ab91b8d2-32c9-4542-a730-7fc84e9892bd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.912559 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82600d19-b64e-4737-8a1f-ff1289daa3ed-kube-api-access-jsw75" (OuterVolumeSpecName: "kube-api-access-jsw75") pod "82600d19-b64e-4737-8a1f-ff1289daa3ed" (UID: "82600d19-b64e-4737-8a1f-ff1289daa3ed"). InnerVolumeSpecName "kube-api-access-jsw75". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.915462 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-scripts" (OuterVolumeSpecName: "scripts") pod "ab91b8d2-32c9-4542-a730-7fc84e9892bd" (UID: "ab91b8d2-32c9-4542-a730-7fc84e9892bd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.924562 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab91b8d2-32c9-4542-a730-7fc84e9892bd-kube-api-access-4msqv" (OuterVolumeSpecName: "kube-api-access-4msqv") pod "ab91b8d2-32c9-4542-a730-7fc84e9892bd" (UID: "ab91b8d2-32c9-4542-a730-7fc84e9892bd"). InnerVolumeSpecName "kube-api-access-4msqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.931256 4775 scope.go:117] "RemoveContainer" containerID="8cc8d28938a70addf2d6412cf322bdf56956d1adec39274fbeadbaacd685266d" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.952404 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82600d19-b64e-4737-8a1f-ff1289daa3ed-config-data" (OuterVolumeSpecName: "config-data") pod "82600d19-b64e-4737-8a1f-ff1289daa3ed" (UID: "82600d19-b64e-4737-8a1f-ff1289daa3ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.959362 4775 scope.go:117] "RemoveContainer" containerID="f3b6a0d856af07fa3386647a1df14c43107fa18d16407653836fe8f6805030db" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.960608 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82600d19-b64e-4737-8a1f-ff1289daa3ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82600d19-b64e-4737-8a1f-ff1289daa3ed" (UID: "82600d19-b64e-4737-8a1f-ff1289daa3ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.961156 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ab91b8d2-32c9-4542-a730-7fc84e9892bd" (UID: "ab91b8d2-32c9-4542-a730-7fc84e9892bd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:19 crc kubenswrapper[4775]: I0321 05:09:19.976299 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82600d19-b64e-4737-8a1f-ff1289daa3ed-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "82600d19-b64e-4737-8a1f-ff1289daa3ed" (UID: "82600d19-b64e-4737-8a1f-ff1289daa3ed"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.008702 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.008730 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82600d19-b64e-4737-8a1f-ff1289daa3ed-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.008740 4775 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82600d19-b64e-4737-8a1f-ff1289daa3ed-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.008751 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82600d19-b64e-4737-8a1f-ff1289daa3ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.008758 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab91b8d2-32c9-4542-a730-7fc84e9892bd-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.008768 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab91b8d2-32c9-4542-a730-7fc84e9892bd-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.008777 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4msqv\" (UniqueName: \"kubernetes.io/projected/ab91b8d2-32c9-4542-a730-7fc84e9892bd-kube-api-access-4msqv\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.008785 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsw75\" (UniqueName: \"kubernetes.io/projected/82600d19-b64e-4737-8a1f-ff1289daa3ed-kube-api-access-jsw75\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.008793 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.034294 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-config-data" (OuterVolumeSpecName: "config-data") pod "ab91b8d2-32c9-4542-a730-7fc84e9892bd" (UID: "ab91b8d2-32c9-4542-a730-7fc84e9892bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.038358 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab91b8d2-32c9-4542-a730-7fc84e9892bd" (UID: "ab91b8d2-32c9-4542-a730-7fc84e9892bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.114082 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.114144 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab91b8d2-32c9-4542-a730-7fc84e9892bd-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.166225 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.178753 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.202078 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.214074 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.225535 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:09:20 crc kubenswrapper[4775]: E0321 05:09:20.227177 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerName="ceilometer-notification-agent" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.227259 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerName="ceilometer-notification-agent" Mar 21 05:09:20 crc kubenswrapper[4775]: E0321 05:09:20.227337 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerName="ceilometer-central-agent" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.227395 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerName="ceilometer-central-agent" Mar 21 05:09:20 crc kubenswrapper[4775]: E0321 05:09:20.227452 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82600d19-b64e-4737-8a1f-ff1289daa3ed" containerName="nova-metadata-metadata" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.227507 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="82600d19-b64e-4737-8a1f-ff1289daa3ed" containerName="nova-metadata-metadata" Mar 21 05:09:20 crc kubenswrapper[4775]: E0321 05:09:20.227568 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerName="proxy-httpd" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.227620 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerName="proxy-httpd" Mar 21 05:09:20 crc kubenswrapper[4775]: E0321 05:09:20.227672 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a407dd6-593f-4806-884b-8e031639d25d" containerName="nova-manage" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.227725 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a407dd6-593f-4806-884b-8e031639d25d" containerName="nova-manage" Mar 21 05:09:20 crc kubenswrapper[4775]: E0321 05:09:20.227797 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7dc929b-501f-43cb-8811-394841cc54f9" containerName="init" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.227863 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7dc929b-501f-43cb-8811-394841cc54f9" containerName="init" Mar 21 05:09:20 crc kubenswrapper[4775]: E0321 05:09:20.227943 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerName="sg-core" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.228008 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerName="sg-core" Mar 21 05:09:20 crc kubenswrapper[4775]: E0321 05:09:20.228072 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7dc929b-501f-43cb-8811-394841cc54f9" containerName="dnsmasq-dns" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.228177 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7dc929b-501f-43cb-8811-394841cc54f9" containerName="dnsmasq-dns" Mar 21 05:09:20 crc kubenswrapper[4775]: E0321 05:09:20.228242 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82600d19-b64e-4737-8a1f-ff1289daa3ed" containerName="nova-metadata-log" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.228295 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="82600d19-b64e-4737-8a1f-ff1289daa3ed" containerName="nova-metadata-log" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.228528 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerName="ceilometer-central-agent" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.228602 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="82600d19-b64e-4737-8a1f-ff1289daa3ed" containerName="nova-metadata-metadata" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.228661 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerName="proxy-httpd" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.228727 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a407dd6-593f-4806-884b-8e031639d25d" containerName="nova-manage" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.228785 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="82600d19-b64e-4737-8a1f-ff1289daa3ed" containerName="nova-metadata-log" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.228839 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerName="ceilometer-notification-agent" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.228893 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7dc929b-501f-43cb-8811-394841cc54f9" containerName="dnsmasq-dns" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.228942 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" containerName="sg-core" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.230948 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.236541 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.236759 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.240272 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.253574 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.257035 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.257366 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.285442 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.292831 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.316514 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/636228b7-669b-4b5d-abec-bf78cb1513f0-logs\") pod \"nova-metadata-0\" (UID: \"636228b7-669b-4b5d-abec-bf78cb1513f0\") " pod="openstack/nova-metadata-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.316569 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-config-data\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.316666 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-log-httpd\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.316701 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.316875 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-run-httpd\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.316921 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.317007 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/636228b7-669b-4b5d-abec-bf78cb1513f0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"636228b7-669b-4b5d-abec-bf78cb1513f0\") " pod="openstack/nova-metadata-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.317146 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636228b7-669b-4b5d-abec-bf78cb1513f0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"636228b7-669b-4b5d-abec-bf78cb1513f0\") " pod="openstack/nova-metadata-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.317180 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-scripts\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.317224 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plq2h\" (UniqueName: \"kubernetes.io/projected/636228b7-669b-4b5d-abec-bf78cb1513f0-kube-api-access-plq2h\") pod \"nova-metadata-0\" (UID: \"636228b7-669b-4b5d-abec-bf78cb1513f0\") " pod="openstack/nova-metadata-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.317262 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsrdg\" (UniqueName: \"kubernetes.io/projected/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-kube-api-access-zsrdg\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.317287 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636228b7-669b-4b5d-abec-bf78cb1513f0-config-data\") pod \"nova-metadata-0\" (UID: \"636228b7-669b-4b5d-abec-bf78cb1513f0\") " pod="openstack/nova-metadata-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.419199 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-run-httpd\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.419261 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.419305 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/636228b7-669b-4b5d-abec-bf78cb1513f0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"636228b7-669b-4b5d-abec-bf78cb1513f0\") " pod="openstack/nova-metadata-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.419343 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636228b7-669b-4b5d-abec-bf78cb1513f0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"636228b7-669b-4b5d-abec-bf78cb1513f0\") " pod="openstack/nova-metadata-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.419393 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-scripts\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.419425 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plq2h\" (UniqueName: \"kubernetes.io/projected/636228b7-669b-4b5d-abec-bf78cb1513f0-kube-api-access-plq2h\") pod \"nova-metadata-0\" (UID: \"636228b7-669b-4b5d-abec-bf78cb1513f0\") " pod="openstack/nova-metadata-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.419449 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsrdg\" (UniqueName: \"kubernetes.io/projected/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-kube-api-access-zsrdg\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.419467 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636228b7-669b-4b5d-abec-bf78cb1513f0-config-data\") pod \"nova-metadata-0\" (UID: \"636228b7-669b-4b5d-abec-bf78cb1513f0\") " pod="openstack/nova-metadata-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.419503 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/636228b7-669b-4b5d-abec-bf78cb1513f0-logs\") pod \"nova-metadata-0\" (UID: \"636228b7-669b-4b5d-abec-bf78cb1513f0\") " pod="openstack/nova-metadata-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.419521 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-config-data\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.419537 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-log-httpd\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.419568 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.419670 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-run-httpd\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.420162 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/636228b7-669b-4b5d-abec-bf78cb1513f0-logs\") pod \"nova-metadata-0\" (UID: \"636228b7-669b-4b5d-abec-bf78cb1513f0\") " pod="openstack/nova-metadata-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.420788 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-log-httpd\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.422488 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/636228b7-669b-4b5d-abec-bf78cb1513f0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"636228b7-669b-4b5d-abec-bf78cb1513f0\") " pod="openstack/nova-metadata-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.423090 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.428263 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.428804 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-scripts\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.434101 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-config-data\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.444738 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636228b7-669b-4b5d-abec-bf78cb1513f0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"636228b7-669b-4b5d-abec-bf78cb1513f0\") " pod="openstack/nova-metadata-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.454761 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plq2h\" (UniqueName: \"kubernetes.io/projected/636228b7-669b-4b5d-abec-bf78cb1513f0-kube-api-access-plq2h\") pod \"nova-metadata-0\" (UID: \"636228b7-669b-4b5d-abec-bf78cb1513f0\") " pod="openstack/nova-metadata-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.455063 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636228b7-669b-4b5d-abec-bf78cb1513f0-config-data\") pod \"nova-metadata-0\" (UID: \"636228b7-669b-4b5d-abec-bf78cb1513f0\") " pod="openstack/nova-metadata-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.465061 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsrdg\" (UniqueName: \"kubernetes.io/projected/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-kube-api-access-zsrdg\") pod \"ceilometer-0\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.567014 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.582839 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 05:09:20 crc kubenswrapper[4775]: I0321 05:09:20.841010 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="e114fa87-9e65-4a0b-bad1-81f521aeef85" containerName="nova-scheduler-scheduler" containerID="cri-o://4db8a2108514793dd399803fe0f901aa4854b0ae09ceaf27535063e2f431666a" gracePeriod=30 Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.081459 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.197796 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:09:21 crc kubenswrapper[4775]: W0321 05:09:21.201032 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod636228b7_669b_4b5d_abec_bf78cb1513f0.slice/crio-c6bcff38ed2ee79c0028a43362b2200b49c362342fd98956493476b35ddff899 WatchSource:0}: Error finding container c6bcff38ed2ee79c0028a43362b2200b49c362342fd98956493476b35ddff899: Status 404 returned error can't find the container with id c6bcff38ed2ee79c0028a43362b2200b49c362342fd98956493476b35ddff899 Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.338212 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vd764" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.372460 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fmmn\" (UniqueName: \"kubernetes.io/projected/fb1dd6e9-8801-464e-952c-d345498b132a-kube-api-access-5fmmn\") pod \"fb1dd6e9-8801-464e-952c-d345498b132a\" (UID: \"fb1dd6e9-8801-464e-952c-d345498b132a\") " Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.372593 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb1dd6e9-8801-464e-952c-d345498b132a-scripts\") pod \"fb1dd6e9-8801-464e-952c-d345498b132a\" (UID: \"fb1dd6e9-8801-464e-952c-d345498b132a\") " Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.372643 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb1dd6e9-8801-464e-952c-d345498b132a-config-data\") pod \"fb1dd6e9-8801-464e-952c-d345498b132a\" (UID: \"fb1dd6e9-8801-464e-952c-d345498b132a\") " Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.372668 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb1dd6e9-8801-464e-952c-d345498b132a-combined-ca-bundle\") pod \"fb1dd6e9-8801-464e-952c-d345498b132a\" (UID: \"fb1dd6e9-8801-464e-952c-d345498b132a\") " Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.384809 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb1dd6e9-8801-464e-952c-d345498b132a-scripts" (OuterVolumeSpecName: "scripts") pod "fb1dd6e9-8801-464e-952c-d345498b132a" (UID: "fb1dd6e9-8801-464e-952c-d345498b132a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.387685 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb1dd6e9-8801-464e-952c-d345498b132a-kube-api-access-5fmmn" (OuterVolumeSpecName: "kube-api-access-5fmmn") pod "fb1dd6e9-8801-464e-952c-d345498b132a" (UID: "fb1dd6e9-8801-464e-952c-d345498b132a"). InnerVolumeSpecName "kube-api-access-5fmmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.436149 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb1dd6e9-8801-464e-952c-d345498b132a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb1dd6e9-8801-464e-952c-d345498b132a" (UID: "fb1dd6e9-8801-464e-952c-d345498b132a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.438990 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb1dd6e9-8801-464e-952c-d345498b132a-config-data" (OuterVolumeSpecName: "config-data") pod "fb1dd6e9-8801-464e-952c-d345498b132a" (UID: "fb1dd6e9-8801-464e-952c-d345498b132a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.474472 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fmmn\" (UniqueName: \"kubernetes.io/projected/fb1dd6e9-8801-464e-952c-d345498b132a-kube-api-access-5fmmn\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.474499 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb1dd6e9-8801-464e-952c-d345498b132a-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.474507 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb1dd6e9-8801-464e-952c-d345498b132a-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.474519 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb1dd6e9-8801-464e-952c-d345498b132a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.672954 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82600d19-b64e-4737-8a1f-ff1289daa3ed" path="/var/lib/kubelet/pods/82600d19-b64e-4737-8a1f-ff1289daa3ed/volumes" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.673569 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab91b8d2-32c9-4542-a730-7fc84e9892bd" path="/var/lib/kubelet/pods/ab91b8d2-32c9-4542-a730-7fc84e9892bd/volumes" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.852984 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"636228b7-669b-4b5d-abec-bf78cb1513f0","Type":"ContainerStarted","Data":"88b942c0d478eda3ab139d99845879b76d209947d510208b64a88f8f909afefc"} Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.853033 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"636228b7-669b-4b5d-abec-bf78cb1513f0","Type":"ContainerStarted","Data":"9a436c521b9cde86255d5d1bb86e7e7617657d67544c27a71666b10506ae93dc"} Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.853048 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"636228b7-669b-4b5d-abec-bf78cb1513f0","Type":"ContainerStarted","Data":"c6bcff38ed2ee79c0028a43362b2200b49c362342fd98956493476b35ddff899"} Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.855251 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vd764" event={"ID":"fb1dd6e9-8801-464e-952c-d345498b132a","Type":"ContainerDied","Data":"5e30087cecb9bbd6f3df2a364239f70f52ef2de20ec9c4dfd84fb90814b44306"} Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.855299 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e30087cecb9bbd6f3df2a364239f70f52ef2de20ec9c4dfd84fb90814b44306" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.855308 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vd764" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.857613 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53","Type":"ContainerStarted","Data":"fbf261dc152da2105525608f99cd085cb605f4c4def074f0140516ec04df21eb"} Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.857646 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53","Type":"ContainerStarted","Data":"bd088ed9e7536178799aa0168c0a00723a8b5b8b146cbd4aa52d938131eb733b"} Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.873810 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.8737908349999999 podStartE2EDuration="1.873790835s" podCreationTimestamp="2026-03-21 05:09:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:09:21.868640489 +0000 UTC m=+1314.845104113" watchObservedRunningTime="2026-03-21 05:09:21.873790835 +0000 UTC m=+1314.850254459" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.939000 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 21 05:09:21 crc kubenswrapper[4775]: E0321 05:09:21.939743 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb1dd6e9-8801-464e-952c-d345498b132a" containerName="nova-cell1-conductor-db-sync" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.939764 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb1dd6e9-8801-464e-952c-d345498b132a" containerName="nova-cell1-conductor-db-sync" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.939939 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb1dd6e9-8801-464e-952c-d345498b132a" containerName="nova-cell1-conductor-db-sync" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.940560 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.944348 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.950136 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.985218 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b29323a7-476f-4a13-8085-4b2158a68850-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b29323a7-476f-4a13-8085-4b2158a68850\") " pod="openstack/nova-cell1-conductor-0" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.985259 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl24b\" (UniqueName: \"kubernetes.io/projected/b29323a7-476f-4a13-8085-4b2158a68850-kube-api-access-kl24b\") pod \"nova-cell1-conductor-0\" (UID: \"b29323a7-476f-4a13-8085-4b2158a68850\") " pod="openstack/nova-cell1-conductor-0" Mar 21 05:09:21 crc kubenswrapper[4775]: I0321 05:09:21.985292 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b29323a7-476f-4a13-8085-4b2158a68850-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b29323a7-476f-4a13-8085-4b2158a68850\") " pod="openstack/nova-cell1-conductor-0" Mar 21 05:09:22 crc kubenswrapper[4775]: I0321 05:09:22.087555 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b29323a7-476f-4a13-8085-4b2158a68850-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b29323a7-476f-4a13-8085-4b2158a68850\") " pod="openstack/nova-cell1-conductor-0" Mar 21 05:09:22 crc kubenswrapper[4775]: I0321 05:09:22.087617 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl24b\" (UniqueName: \"kubernetes.io/projected/b29323a7-476f-4a13-8085-4b2158a68850-kube-api-access-kl24b\") pod \"nova-cell1-conductor-0\" (UID: \"b29323a7-476f-4a13-8085-4b2158a68850\") " pod="openstack/nova-cell1-conductor-0" Mar 21 05:09:22 crc kubenswrapper[4775]: I0321 05:09:22.087652 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b29323a7-476f-4a13-8085-4b2158a68850-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b29323a7-476f-4a13-8085-4b2158a68850\") " pod="openstack/nova-cell1-conductor-0" Mar 21 05:09:22 crc kubenswrapper[4775]: I0321 05:09:22.094881 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b29323a7-476f-4a13-8085-4b2158a68850-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b29323a7-476f-4a13-8085-4b2158a68850\") " pod="openstack/nova-cell1-conductor-0" Mar 21 05:09:22 crc kubenswrapper[4775]: I0321 05:09:22.094928 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b29323a7-476f-4a13-8085-4b2158a68850-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b29323a7-476f-4a13-8085-4b2158a68850\") " pod="openstack/nova-cell1-conductor-0" Mar 21 05:09:22 crc kubenswrapper[4775]: I0321 05:09:22.107572 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl24b\" (UniqueName: \"kubernetes.io/projected/b29323a7-476f-4a13-8085-4b2158a68850-kube-api-access-kl24b\") pod \"nova-cell1-conductor-0\" (UID: \"b29323a7-476f-4a13-8085-4b2158a68850\") " pod="openstack/nova-cell1-conductor-0" Mar 21 05:09:22 crc kubenswrapper[4775]: I0321 05:09:22.258659 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 21 05:09:22 crc kubenswrapper[4775]: W0321 05:09:22.786145 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb29323a7_476f_4a13_8085_4b2158a68850.slice/crio-00db731407ff126bb6e108665e2bc8cb1ba2533f6f82c1e28ee6f27a51895049 WatchSource:0}: Error finding container 00db731407ff126bb6e108665e2bc8cb1ba2533f6f82c1e28ee6f27a51895049: Status 404 returned error can't find the container with id 00db731407ff126bb6e108665e2bc8cb1ba2533f6f82c1e28ee6f27a51895049 Mar 21 05:09:22 crc kubenswrapper[4775]: I0321 05:09:22.791021 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 21 05:09:22 crc kubenswrapper[4775]: I0321 05:09:22.870968 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b29323a7-476f-4a13-8085-4b2158a68850","Type":"ContainerStarted","Data":"00db731407ff126bb6e108665e2bc8cb1ba2533f6f82c1e28ee6f27a51895049"} Mar 21 05:09:22 crc kubenswrapper[4775]: I0321 05:09:22.874174 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53","Type":"ContainerStarted","Data":"e1897e25daedbe1665e342773e7750ff60df7bc2b0e9065d7163a3b3ead5e27d"} Mar 21 05:09:23 crc kubenswrapper[4775]: E0321 05:09:23.057297 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4db8a2108514793dd399803fe0f901aa4854b0ae09ceaf27535063e2f431666a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 21 05:09:23 crc kubenswrapper[4775]: E0321 05:09:23.058510 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4db8a2108514793dd399803fe0f901aa4854b0ae09ceaf27535063e2f431666a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 21 05:09:23 crc kubenswrapper[4775]: E0321 05:09:23.082518 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4db8a2108514793dd399803fe0f901aa4854b0ae09ceaf27535063e2f431666a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 21 05:09:23 crc kubenswrapper[4775]: E0321 05:09:23.082601 4775 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="e114fa87-9e65-4a0b-bad1-81f521aeef85" containerName="nova-scheduler-scheduler" Mar 21 05:09:23 crc kubenswrapper[4775]: I0321 05:09:23.886816 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b29323a7-476f-4a13-8085-4b2158a68850","Type":"ContainerStarted","Data":"16628276422a39d68f8845f2fe31fdf04cb116e0f714c01cf1ba2a054a43b6d1"} Mar 21 05:09:23 crc kubenswrapper[4775]: I0321 05:09:23.887461 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 21 05:09:23 crc kubenswrapper[4775]: I0321 05:09:23.890499 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53","Type":"ContainerStarted","Data":"8f1be7b1e4c2dbdc319160849ad419dc210086da61fe3aad3c1277cd3f4787db"} Mar 21 05:09:23 crc kubenswrapper[4775]: I0321 05:09:23.919374 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.9193526419999998 podStartE2EDuration="2.919352642s" podCreationTimestamp="2026-03-21 05:09:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:09:23.908525136 +0000 UTC m=+1316.884988770" watchObservedRunningTime="2026-03-21 05:09:23.919352642 +0000 UTC m=+1316.895816286" Mar 21 05:09:24 crc kubenswrapper[4775]: I0321 05:09:24.904105 4775 generic.go:334] "Generic (PLEG): container finished" podID="87dad904-a84b-4dcf-9caa-3fe89053b96b" containerID="3d41a9111ff02a22981deef2e158c7aaae04c761a3f477f9cde2ecff8054c0f1" exitCode=0 Mar 21 05:09:24 crc kubenswrapper[4775]: I0321 05:09:24.904149 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87dad904-a84b-4dcf-9caa-3fe89053b96b","Type":"ContainerDied","Data":"3d41a9111ff02a22981deef2e158c7aaae04c761a3f477f9cde2ecff8054c0f1"} Mar 21 05:09:24 crc kubenswrapper[4775]: I0321 05:09:24.904585 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87dad904-a84b-4dcf-9caa-3fe89053b96b","Type":"ContainerDied","Data":"3c21cc3a28ac294006c3d112d32cbbbabe9d8f787077b90e376062aeec6dab3e"} Mar 21 05:09:24 crc kubenswrapper[4775]: I0321 05:09:24.904605 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c21cc3a28ac294006c3d112d32cbbbabe9d8f787077b90e376062aeec6dab3e" Mar 21 05:09:24 crc kubenswrapper[4775]: I0321 05:09:24.906630 4775 generic.go:334] "Generic (PLEG): container finished" podID="e114fa87-9e65-4a0b-bad1-81f521aeef85" containerID="4db8a2108514793dd399803fe0f901aa4854b0ae09ceaf27535063e2f431666a" exitCode=0 Mar 21 05:09:24 crc kubenswrapper[4775]: I0321 05:09:24.906695 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e114fa87-9e65-4a0b-bad1-81f521aeef85","Type":"ContainerDied","Data":"4db8a2108514793dd399803fe0f901aa4854b0ae09ceaf27535063e2f431666a"} Mar 21 05:09:24 crc kubenswrapper[4775]: I0321 05:09:24.999819 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.036373 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87dad904-a84b-4dcf-9caa-3fe89053b96b-logs\") pod \"87dad904-a84b-4dcf-9caa-3fe89053b96b\" (UID: \"87dad904-a84b-4dcf-9caa-3fe89053b96b\") " Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.036460 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txpj5\" (UniqueName: \"kubernetes.io/projected/87dad904-a84b-4dcf-9caa-3fe89053b96b-kube-api-access-txpj5\") pod \"87dad904-a84b-4dcf-9caa-3fe89053b96b\" (UID: \"87dad904-a84b-4dcf-9caa-3fe89053b96b\") " Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.036502 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87dad904-a84b-4dcf-9caa-3fe89053b96b-config-data\") pod \"87dad904-a84b-4dcf-9caa-3fe89053b96b\" (UID: \"87dad904-a84b-4dcf-9caa-3fe89053b96b\") " Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.036539 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87dad904-a84b-4dcf-9caa-3fe89053b96b-combined-ca-bundle\") pod \"87dad904-a84b-4dcf-9caa-3fe89053b96b\" (UID: \"87dad904-a84b-4dcf-9caa-3fe89053b96b\") " Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.036887 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87dad904-a84b-4dcf-9caa-3fe89053b96b-logs" (OuterVolumeSpecName: "logs") pod "87dad904-a84b-4dcf-9caa-3fe89053b96b" (UID: "87dad904-a84b-4dcf-9caa-3fe89053b96b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.044356 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87dad904-a84b-4dcf-9caa-3fe89053b96b-kube-api-access-txpj5" (OuterVolumeSpecName: "kube-api-access-txpj5") pod "87dad904-a84b-4dcf-9caa-3fe89053b96b" (UID: "87dad904-a84b-4dcf-9caa-3fe89053b96b"). InnerVolumeSpecName "kube-api-access-txpj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.110332 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87dad904-a84b-4dcf-9caa-3fe89053b96b-config-data" (OuterVolumeSpecName: "config-data") pod "87dad904-a84b-4dcf-9caa-3fe89053b96b" (UID: "87dad904-a84b-4dcf-9caa-3fe89053b96b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.114490 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87dad904-a84b-4dcf-9caa-3fe89053b96b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87dad904-a84b-4dcf-9caa-3fe89053b96b" (UID: "87dad904-a84b-4dcf-9caa-3fe89053b96b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.137708 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87dad904-a84b-4dcf-9caa-3fe89053b96b-logs\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.137730 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txpj5\" (UniqueName: \"kubernetes.io/projected/87dad904-a84b-4dcf-9caa-3fe89053b96b-kube-api-access-txpj5\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.137740 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87dad904-a84b-4dcf-9caa-3fe89053b96b-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.137750 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87dad904-a84b-4dcf-9caa-3fe89053b96b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.339375 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.441588 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e114fa87-9e65-4a0b-bad1-81f521aeef85-config-data\") pod \"e114fa87-9e65-4a0b-bad1-81f521aeef85\" (UID: \"e114fa87-9e65-4a0b-bad1-81f521aeef85\") " Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.441648 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e114fa87-9e65-4a0b-bad1-81f521aeef85-combined-ca-bundle\") pod \"e114fa87-9e65-4a0b-bad1-81f521aeef85\" (UID: \"e114fa87-9e65-4a0b-bad1-81f521aeef85\") " Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.441819 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpl5g\" (UniqueName: \"kubernetes.io/projected/e114fa87-9e65-4a0b-bad1-81f521aeef85-kube-api-access-xpl5g\") pod \"e114fa87-9e65-4a0b-bad1-81f521aeef85\" (UID: \"e114fa87-9e65-4a0b-bad1-81f521aeef85\") " Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.445130 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e114fa87-9e65-4a0b-bad1-81f521aeef85-kube-api-access-xpl5g" (OuterVolumeSpecName: "kube-api-access-xpl5g") pod "e114fa87-9e65-4a0b-bad1-81f521aeef85" (UID: "e114fa87-9e65-4a0b-bad1-81f521aeef85"). InnerVolumeSpecName "kube-api-access-xpl5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.484773 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e114fa87-9e65-4a0b-bad1-81f521aeef85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e114fa87-9e65-4a0b-bad1-81f521aeef85" (UID: "e114fa87-9e65-4a0b-bad1-81f521aeef85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.516172 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e114fa87-9e65-4a0b-bad1-81f521aeef85-config-data" (OuterVolumeSpecName: "config-data") pod "e114fa87-9e65-4a0b-bad1-81f521aeef85" (UID: "e114fa87-9e65-4a0b-bad1-81f521aeef85"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.547062 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpl5g\" (UniqueName: \"kubernetes.io/projected/e114fa87-9e65-4a0b-bad1-81f521aeef85-kube-api-access-xpl5g\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.547420 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e114fa87-9e65-4a0b-bad1-81f521aeef85-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.547433 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e114fa87-9e65-4a0b-bad1-81f521aeef85-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.928612 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53","Type":"ContainerStarted","Data":"1019edd2ab7e9cd0ed64f98b04c2e3b4282ef0e3d503d9b2e72d9464a7033ab9"} Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.929369 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.931140 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.931860 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.932232 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e114fa87-9e65-4a0b-bad1-81f521aeef85","Type":"ContainerDied","Data":"8a454c45ca71508ee3e69e89e31c30561bb964893c09c7b3c20372f820d152ee"} Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.932268 4775 scope.go:117] "RemoveContainer" containerID="4db8a2108514793dd399803fe0f901aa4854b0ae09ceaf27535063e2f431666a" Mar 21 05:09:25 crc kubenswrapper[4775]: I0321 05:09:25.950234 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.604476482 podStartE2EDuration="5.950212725s" podCreationTimestamp="2026-03-21 05:09:20 +0000 UTC" firstStartedPulling="2026-03-21 05:09:21.095857523 +0000 UTC m=+1314.072321147" lastFinishedPulling="2026-03-21 05:09:25.441593766 +0000 UTC m=+1318.418057390" observedRunningTime="2026-03-21 05:09:25.947934851 +0000 UTC m=+1318.924398475" watchObservedRunningTime="2026-03-21 05:09:25.950212725 +0000 UTC m=+1318.926676349" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.008863 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.020108 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.044110 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.053235 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 05:09:26 crc kubenswrapper[4775]: E0321 05:09:26.053929 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87dad904-a84b-4dcf-9caa-3fe89053b96b" containerName="nova-api-log" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.053954 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="87dad904-a84b-4dcf-9caa-3fe89053b96b" containerName="nova-api-log" Mar 21 05:09:26 crc kubenswrapper[4775]: E0321 05:09:26.053988 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87dad904-a84b-4dcf-9caa-3fe89053b96b" containerName="nova-api-api" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.053997 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="87dad904-a84b-4dcf-9caa-3fe89053b96b" containerName="nova-api-api" Mar 21 05:09:26 crc kubenswrapper[4775]: E0321 05:09:26.054014 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e114fa87-9e65-4a0b-bad1-81f521aeef85" containerName="nova-scheduler-scheduler" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.054020 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e114fa87-9e65-4a0b-bad1-81f521aeef85" containerName="nova-scheduler-scheduler" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.054248 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e114fa87-9e65-4a0b-bad1-81f521aeef85" containerName="nova-scheduler-scheduler" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.054268 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="87dad904-a84b-4dcf-9caa-3fe89053b96b" containerName="nova-api-log" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.054285 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="87dad904-a84b-4dcf-9caa-3fe89053b96b" containerName="nova-api-api" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.054867 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.060037 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.071928 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.082005 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.089452 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.091027 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.092586 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.099920 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.160537 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f713501-fa26-4ef2-87ac-c5376b84891d-config-data\") pod \"nova-scheduler-0\" (UID: \"0f713501-fa26-4ef2-87ac-c5376b84891d\") " pod="openstack/nova-scheduler-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.160640 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f713501-fa26-4ef2-87ac-c5376b84891d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0f713501-fa26-4ef2-87ac-c5376b84891d\") " pod="openstack/nova-scheduler-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.160665 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d4895fa-5b3e-423f-85aa-030d2db076b3-logs\") pod \"nova-api-0\" (UID: \"7d4895fa-5b3e-423f-85aa-030d2db076b3\") " pod="openstack/nova-api-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.160765 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krq52\" (UniqueName: \"kubernetes.io/projected/7d4895fa-5b3e-423f-85aa-030d2db076b3-kube-api-access-krq52\") pod \"nova-api-0\" (UID: \"7d4895fa-5b3e-423f-85aa-030d2db076b3\") " pod="openstack/nova-api-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.160795 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d4895fa-5b3e-423f-85aa-030d2db076b3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7d4895fa-5b3e-423f-85aa-030d2db076b3\") " pod="openstack/nova-api-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.160812 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d4895fa-5b3e-423f-85aa-030d2db076b3-config-data\") pod \"nova-api-0\" (UID: \"7d4895fa-5b3e-423f-85aa-030d2db076b3\") " pod="openstack/nova-api-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.160844 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jkv8\" (UniqueName: \"kubernetes.io/projected/0f713501-fa26-4ef2-87ac-c5376b84891d-kube-api-access-5jkv8\") pod \"nova-scheduler-0\" (UID: \"0f713501-fa26-4ef2-87ac-c5376b84891d\") " pod="openstack/nova-scheduler-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.262787 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d4895fa-5b3e-423f-85aa-030d2db076b3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7d4895fa-5b3e-423f-85aa-030d2db076b3\") " pod="openstack/nova-api-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.262845 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d4895fa-5b3e-423f-85aa-030d2db076b3-config-data\") pod \"nova-api-0\" (UID: \"7d4895fa-5b3e-423f-85aa-030d2db076b3\") " pod="openstack/nova-api-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.262888 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jkv8\" (UniqueName: \"kubernetes.io/projected/0f713501-fa26-4ef2-87ac-c5376b84891d-kube-api-access-5jkv8\") pod \"nova-scheduler-0\" (UID: \"0f713501-fa26-4ef2-87ac-c5376b84891d\") " pod="openstack/nova-scheduler-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.262965 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f713501-fa26-4ef2-87ac-c5376b84891d-config-data\") pod \"nova-scheduler-0\" (UID: \"0f713501-fa26-4ef2-87ac-c5376b84891d\") " pod="openstack/nova-scheduler-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.263010 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f713501-fa26-4ef2-87ac-c5376b84891d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0f713501-fa26-4ef2-87ac-c5376b84891d\") " pod="openstack/nova-scheduler-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.263032 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d4895fa-5b3e-423f-85aa-030d2db076b3-logs\") pod \"nova-api-0\" (UID: \"7d4895fa-5b3e-423f-85aa-030d2db076b3\") " pod="openstack/nova-api-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.263065 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krq52\" (UniqueName: \"kubernetes.io/projected/7d4895fa-5b3e-423f-85aa-030d2db076b3-kube-api-access-krq52\") pod \"nova-api-0\" (UID: \"7d4895fa-5b3e-423f-85aa-030d2db076b3\") " pod="openstack/nova-api-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.264389 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d4895fa-5b3e-423f-85aa-030d2db076b3-logs\") pod \"nova-api-0\" (UID: \"7d4895fa-5b3e-423f-85aa-030d2db076b3\") " pod="openstack/nova-api-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.268096 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f713501-fa26-4ef2-87ac-c5376b84891d-config-data\") pod \"nova-scheduler-0\" (UID: \"0f713501-fa26-4ef2-87ac-c5376b84891d\") " pod="openstack/nova-scheduler-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.270268 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f713501-fa26-4ef2-87ac-c5376b84891d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0f713501-fa26-4ef2-87ac-c5376b84891d\") " pod="openstack/nova-scheduler-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.282892 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d4895fa-5b3e-423f-85aa-030d2db076b3-config-data\") pod \"nova-api-0\" (UID: \"7d4895fa-5b3e-423f-85aa-030d2db076b3\") " pod="openstack/nova-api-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.285098 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d4895fa-5b3e-423f-85aa-030d2db076b3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7d4895fa-5b3e-423f-85aa-030d2db076b3\") " pod="openstack/nova-api-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.285830 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jkv8\" (UniqueName: \"kubernetes.io/projected/0f713501-fa26-4ef2-87ac-c5376b84891d-kube-api-access-5jkv8\") pod \"nova-scheduler-0\" (UID: \"0f713501-fa26-4ef2-87ac-c5376b84891d\") " pod="openstack/nova-scheduler-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.286338 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krq52\" (UniqueName: \"kubernetes.io/projected/7d4895fa-5b3e-423f-85aa-030d2db076b3-kube-api-access-krq52\") pod \"nova-api-0\" (UID: \"7d4895fa-5b3e-423f-85aa-030d2db076b3\") " pod="openstack/nova-api-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.372186 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.416393 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.855929 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 05:09:26 crc kubenswrapper[4775]: W0321 05:09:26.866073 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f713501_fa26_4ef2_87ac_c5376b84891d.slice/crio-35d35b17b5e3ddd8509d6d47f432a3246971c8785668db26e3bf835f5d64e720 WatchSource:0}: Error finding container 35d35b17b5e3ddd8509d6d47f432a3246971c8785668db26e3bf835f5d64e720: Status 404 returned error can't find the container with id 35d35b17b5e3ddd8509d6d47f432a3246971c8785668db26e3bf835f5d64e720 Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.922150 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 21 05:09:26 crc kubenswrapper[4775]: W0321 05:09:26.928564 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d4895fa_5b3e_423f_85aa_030d2db076b3.slice/crio-d95657026d4984394319083a88a788fdbce021127cb6cbf909a8f8735dc34db8 WatchSource:0}: Error finding container d95657026d4984394319083a88a788fdbce021127cb6cbf909a8f8735dc34db8: Status 404 returned error can't find the container with id d95657026d4984394319083a88a788fdbce021127cb6cbf909a8f8735dc34db8 Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.947346 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7d4895fa-5b3e-423f-85aa-030d2db076b3","Type":"ContainerStarted","Data":"d95657026d4984394319083a88a788fdbce021127cb6cbf909a8f8735dc34db8"} Mar 21 05:09:26 crc kubenswrapper[4775]: I0321 05:09:26.950856 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0f713501-fa26-4ef2-87ac-c5376b84891d","Type":"ContainerStarted","Data":"35d35b17b5e3ddd8509d6d47f432a3246971c8785668db26e3bf835f5d64e720"} Mar 21 05:09:27 crc kubenswrapper[4775]: I0321 05:09:27.306040 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 21 05:09:27 crc kubenswrapper[4775]: I0321 05:09:27.702908 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87dad904-a84b-4dcf-9caa-3fe89053b96b" path="/var/lib/kubelet/pods/87dad904-a84b-4dcf-9caa-3fe89053b96b/volumes" Mar 21 05:09:27 crc kubenswrapper[4775]: I0321 05:09:27.703822 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e114fa87-9e65-4a0b-bad1-81f521aeef85" path="/var/lib/kubelet/pods/e114fa87-9e65-4a0b-bad1-81f521aeef85/volumes" Mar 21 05:09:27 crc kubenswrapper[4775]: I0321 05:09:27.982368 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7d4895fa-5b3e-423f-85aa-030d2db076b3","Type":"ContainerStarted","Data":"ec80243e9f9293a7a7355096362bcfe00d0fa0d7ef991f719431c39710253285"} Mar 21 05:09:27 crc kubenswrapper[4775]: I0321 05:09:27.982423 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7d4895fa-5b3e-423f-85aa-030d2db076b3","Type":"ContainerStarted","Data":"ad57dd190fc32b9e6c21bd060d276edc5365c62c5bf0f9ffb91fab933879f448"} Mar 21 05:09:27 crc kubenswrapper[4775]: I0321 05:09:27.984893 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0f713501-fa26-4ef2-87ac-c5376b84891d","Type":"ContainerStarted","Data":"c561875841331ee73e6bcb2e365ce80407bc0539d3ec5ad570942023fedeb9c2"} Mar 21 05:09:28 crc kubenswrapper[4775]: I0321 05:09:28.019712 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.019692629 podStartE2EDuration="3.019692629s" podCreationTimestamp="2026-03-21 05:09:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:09:28.012452044 +0000 UTC m=+1320.988915678" watchObservedRunningTime="2026-03-21 05:09:28.019692629 +0000 UTC m=+1320.996156253" Mar 21 05:09:28 crc kubenswrapper[4775]: I0321 05:09:28.030097 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.030077242 podStartE2EDuration="3.030077242s" podCreationTimestamp="2026-03-21 05:09:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:09:28.02857394 +0000 UTC m=+1321.005037564" watchObservedRunningTime="2026-03-21 05:09:28.030077242 +0000 UTC m=+1321.006540866" Mar 21 05:09:30 crc kubenswrapper[4775]: I0321 05:09:30.583692 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 21 05:09:30 crc kubenswrapper[4775]: I0321 05:09:30.584088 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 21 05:09:31 crc kubenswrapper[4775]: I0321 05:09:31.375338 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 21 05:09:31 crc kubenswrapper[4775]: I0321 05:09:31.600478 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="636228b7-669b-4b5d-abec-bf78cb1513f0" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 21 05:09:31 crc kubenswrapper[4775]: I0321 05:09:31.600463 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="636228b7-669b-4b5d-abec-bf78cb1513f0" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 21 05:09:36 crc kubenswrapper[4775]: I0321 05:09:36.373810 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 21 05:09:36 crc kubenswrapper[4775]: I0321 05:09:36.417050 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 21 05:09:36 crc kubenswrapper[4775]: I0321 05:09:36.417108 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 21 05:09:36 crc kubenswrapper[4775]: I0321 05:09:36.419536 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 21 05:09:37 crc kubenswrapper[4775]: I0321 05:09:37.094500 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 21 05:09:37 crc kubenswrapper[4775]: I0321 05:09:37.499447 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7d4895fa-5b3e-423f-85aa-030d2db076b3" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 21 05:09:37 crc kubenswrapper[4775]: I0321 05:09:37.499466 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7d4895fa-5b3e-423f-85aa-030d2db076b3" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 21 05:09:38 crc kubenswrapper[4775]: I0321 05:09:38.583208 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 21 05:09:38 crc kubenswrapper[4775]: I0321 05:09:38.583270 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 21 05:09:40 crc kubenswrapper[4775]: I0321 05:09:40.589501 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 21 05:09:40 crc kubenswrapper[4775]: I0321 05:09:40.591587 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 21 05:09:40 crc kubenswrapper[4775]: I0321 05:09:40.598573 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 21 05:09:41 crc kubenswrapper[4775]: I0321 05:09:41.107517 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.028306 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.118282 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2-config-data\") pod \"e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2\" (UID: \"e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2\") " Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.118409 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2-combined-ca-bundle\") pod \"e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2\" (UID: \"e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2\") " Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.118793 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swsh9\" (UniqueName: \"kubernetes.io/projected/e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2-kube-api-access-swsh9\") pod \"e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2\" (UID: \"e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2\") " Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.126031 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2-kube-api-access-swsh9" (OuterVolumeSpecName: "kube-api-access-swsh9") pod "e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2" (UID: "e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2"). InnerVolumeSpecName "kube-api-access-swsh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.138608 4775 generic.go:334] "Generic (PLEG): container finished" podID="e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2" containerID="bdf97a0cf8b488689d63ad976ec7a07bd6c1bed2c6f6a773ddf11243bbcc10aa" exitCode=137 Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.138693 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.138754 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2","Type":"ContainerDied","Data":"bdf97a0cf8b488689d63ad976ec7a07bd6c1bed2c6f6a773ddf11243bbcc10aa"} Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.139266 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2","Type":"ContainerDied","Data":"e7ffc178120764777a6e0c0217aa92bdaf948da3b3e7aaf4502f70ada80abfb2"} Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.139307 4775 scope.go:117] "RemoveContainer" containerID="bdf97a0cf8b488689d63ad976ec7a07bd6c1bed2c6f6a773ddf11243bbcc10aa" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.150616 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2-config-data" (OuterVolumeSpecName: "config-data") pod "e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2" (UID: "e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.162088 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2" (UID: "e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.221639 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.221690 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.221702 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swsh9\" (UniqueName: \"kubernetes.io/projected/e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2-kube-api-access-swsh9\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.224006 4775 scope.go:117] "RemoveContainer" containerID="bdf97a0cf8b488689d63ad976ec7a07bd6c1bed2c6f6a773ddf11243bbcc10aa" Mar 21 05:09:44 crc kubenswrapper[4775]: E0321 05:09:44.224510 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdf97a0cf8b488689d63ad976ec7a07bd6c1bed2c6f6a773ddf11243bbcc10aa\": container with ID starting with bdf97a0cf8b488689d63ad976ec7a07bd6c1bed2c6f6a773ddf11243bbcc10aa not found: ID does not exist" containerID="bdf97a0cf8b488689d63ad976ec7a07bd6c1bed2c6f6a773ddf11243bbcc10aa" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.224544 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdf97a0cf8b488689d63ad976ec7a07bd6c1bed2c6f6a773ddf11243bbcc10aa"} err="failed to get container status \"bdf97a0cf8b488689d63ad976ec7a07bd6c1bed2c6f6a773ddf11243bbcc10aa\": rpc error: code = NotFound desc = could not find container \"bdf97a0cf8b488689d63ad976ec7a07bd6c1bed2c6f6a773ddf11243bbcc10aa\": container with ID starting with bdf97a0cf8b488689d63ad976ec7a07bd6c1bed2c6f6a773ddf11243bbcc10aa not found: ID does not exist" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.417528 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.417825 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.487500 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.501236 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.524797 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 21 05:09:44 crc kubenswrapper[4775]: E0321 05:09:44.525425 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2" containerName="nova-cell1-novncproxy-novncproxy" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.525456 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2" containerName="nova-cell1-novncproxy-novncproxy" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.525743 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2" containerName="nova-cell1-novncproxy-novncproxy" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.526696 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.529067 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.529987 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.530298 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.538649 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.627278 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx5bx\" (UniqueName: \"kubernetes.io/projected/3b7ea443-e30d-41d1-9f42-0bef9d7bd012-kube-api-access-wx5bx\") pod \"nova-cell1-novncproxy-0\" (UID: \"3b7ea443-e30d-41d1-9f42-0bef9d7bd012\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.627383 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b7ea443-e30d-41d1-9f42-0bef9d7bd012-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3b7ea443-e30d-41d1-9f42-0bef9d7bd012\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.627519 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b7ea443-e30d-41d1-9f42-0bef9d7bd012-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3b7ea443-e30d-41d1-9f42-0bef9d7bd012\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.627616 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b7ea443-e30d-41d1-9f42-0bef9d7bd012-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3b7ea443-e30d-41d1-9f42-0bef9d7bd012\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.627671 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b7ea443-e30d-41d1-9f42-0bef9d7bd012-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3b7ea443-e30d-41d1-9f42-0bef9d7bd012\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.729750 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b7ea443-e30d-41d1-9f42-0bef9d7bd012-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3b7ea443-e30d-41d1-9f42-0bef9d7bd012\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.729918 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b7ea443-e30d-41d1-9f42-0bef9d7bd012-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3b7ea443-e30d-41d1-9f42-0bef9d7bd012\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.730037 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b7ea443-e30d-41d1-9f42-0bef9d7bd012-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3b7ea443-e30d-41d1-9f42-0bef9d7bd012\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.730114 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b7ea443-e30d-41d1-9f42-0bef9d7bd012-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3b7ea443-e30d-41d1-9f42-0bef9d7bd012\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.730248 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx5bx\" (UniqueName: \"kubernetes.io/projected/3b7ea443-e30d-41d1-9f42-0bef9d7bd012-kube-api-access-wx5bx\") pod \"nova-cell1-novncproxy-0\" (UID: \"3b7ea443-e30d-41d1-9f42-0bef9d7bd012\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.734247 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b7ea443-e30d-41d1-9f42-0bef9d7bd012-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3b7ea443-e30d-41d1-9f42-0bef9d7bd012\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.734478 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b7ea443-e30d-41d1-9f42-0bef9d7bd012-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3b7ea443-e30d-41d1-9f42-0bef9d7bd012\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.734920 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b7ea443-e30d-41d1-9f42-0bef9d7bd012-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3b7ea443-e30d-41d1-9f42-0bef9d7bd012\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.736386 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b7ea443-e30d-41d1-9f42-0bef9d7bd012-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3b7ea443-e30d-41d1-9f42-0bef9d7bd012\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.753150 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx5bx\" (UniqueName: \"kubernetes.io/projected/3b7ea443-e30d-41d1-9f42-0bef9d7bd012-kube-api-access-wx5bx\") pod \"nova-cell1-novncproxy-0\" (UID: \"3b7ea443-e30d-41d1-9f42-0bef9d7bd012\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:44 crc kubenswrapper[4775]: I0321 05:09:44.860132 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:45 crc kubenswrapper[4775]: W0321 05:09:45.330930 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b7ea443_e30d_41d1_9f42_0bef9d7bd012.slice/crio-b222bd3e248b9d0166f90bc7b6192e1078b253859cb36e93ca4b525699197a5f WatchSource:0}: Error finding container b222bd3e248b9d0166f90bc7b6192e1078b253859cb36e93ca4b525699197a5f: Status 404 returned error can't find the container with id b222bd3e248b9d0166f90bc7b6192e1078b253859cb36e93ca4b525699197a5f Mar 21 05:09:45 crc kubenswrapper[4775]: I0321 05:09:45.331983 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 21 05:09:45 crc kubenswrapper[4775]: I0321 05:09:45.680532 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2" path="/var/lib/kubelet/pods/e5cc1cb8-86fa-4708-bd8f-d8da98ebc0b2/volumes" Mar 21 05:09:46 crc kubenswrapper[4775]: I0321 05:09:46.157811 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3b7ea443-e30d-41d1-9f42-0bef9d7bd012","Type":"ContainerStarted","Data":"48d8fa4a899bf40e560d38784c8c256d93eddd6127cafea852c9a1cba4bc7d1b"} Mar 21 05:09:46 crc kubenswrapper[4775]: I0321 05:09:46.158329 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3b7ea443-e30d-41d1-9f42-0bef9d7bd012","Type":"ContainerStarted","Data":"b222bd3e248b9d0166f90bc7b6192e1078b253859cb36e93ca4b525699197a5f"} Mar 21 05:09:46 crc kubenswrapper[4775]: I0321 05:09:46.182192 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.182173831 podStartE2EDuration="2.182173831s" podCreationTimestamp="2026-03-21 05:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:09:46.17542978 +0000 UTC m=+1339.151893404" watchObservedRunningTime="2026-03-21 05:09:46.182173831 +0000 UTC m=+1339.158637455" Mar 21 05:09:46 crc kubenswrapper[4775]: I0321 05:09:46.424755 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 21 05:09:46 crc kubenswrapper[4775]: I0321 05:09:46.425857 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 21 05:09:46 crc kubenswrapper[4775]: I0321 05:09:46.439539 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.168437 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.378897 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-rtnbf"] Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.380394 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.391378 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-rtnbf"] Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.429654 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-rtnbf\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.429752 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl7xb\" (UniqueName: \"kubernetes.io/projected/d725aabc-ba32-4c0e-bc91-8819d73cae40-kube-api-access-dl7xb\") pod \"dnsmasq-dns-cd5cbd7b9-rtnbf\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.429779 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-rtnbf\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.429802 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-config\") pod \"dnsmasq-dns-cd5cbd7b9-rtnbf\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.429879 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-rtnbf\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.429958 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-rtnbf\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.531905 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-rtnbf\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.531995 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-rtnbf\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.532045 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-rtnbf\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.532082 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl7xb\" (UniqueName: \"kubernetes.io/projected/d725aabc-ba32-4c0e-bc91-8819d73cae40-kube-api-access-dl7xb\") pod \"dnsmasq-dns-cd5cbd7b9-rtnbf\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.532106 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-rtnbf\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.532132 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-config\") pod \"dnsmasq-dns-cd5cbd7b9-rtnbf\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.533144 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-config\") pod \"dnsmasq-dns-cd5cbd7b9-rtnbf\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.533451 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-rtnbf\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.533529 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-rtnbf\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.533572 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-rtnbf\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.533875 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-rtnbf\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.559465 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl7xb\" (UniqueName: \"kubernetes.io/projected/d725aabc-ba32-4c0e-bc91-8819d73cae40-kube-api-access-dl7xb\") pod \"dnsmasq-dns-cd5cbd7b9-rtnbf\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:47 crc kubenswrapper[4775]: I0321 05:09:47.740540 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:48 crc kubenswrapper[4775]: W0321 05:09:48.251274 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd725aabc_ba32_4c0e_bc91_8819d73cae40.slice/crio-5dc8047188086fad5c8d0b1fabab1913e5036d9e22f49d173fa13d6dd588e3c0 WatchSource:0}: Error finding container 5dc8047188086fad5c8d0b1fabab1913e5036d9e22f49d173fa13d6dd588e3c0: Status 404 returned error can't find the container with id 5dc8047188086fad5c8d0b1fabab1913e5036d9e22f49d173fa13d6dd588e3c0 Mar 21 05:09:48 crc kubenswrapper[4775]: I0321 05:09:48.255029 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-rtnbf"] Mar 21 05:09:49 crc kubenswrapper[4775]: I0321 05:09:49.184506 4775 generic.go:334] "Generic (PLEG): container finished" podID="d725aabc-ba32-4c0e-bc91-8819d73cae40" containerID="e55382d5f315dca58dbeb0320372028174ccdbb56163037ad99b712395ff7974" exitCode=0 Mar 21 05:09:49 crc kubenswrapper[4775]: I0321 05:09:49.184726 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" event={"ID":"d725aabc-ba32-4c0e-bc91-8819d73cae40","Type":"ContainerDied","Data":"e55382d5f315dca58dbeb0320372028174ccdbb56163037ad99b712395ff7974"} Mar 21 05:09:49 crc kubenswrapper[4775]: I0321 05:09:49.185165 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" event={"ID":"d725aabc-ba32-4c0e-bc91-8819d73cae40","Type":"ContainerStarted","Data":"5dc8047188086fad5c8d0b1fabab1913e5036d9e22f49d173fa13d6dd588e3c0"} Mar 21 05:09:49 crc kubenswrapper[4775]: I0321 05:09:49.673954 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:09:49 crc kubenswrapper[4775]: I0321 05:09:49.674292 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerName="ceilometer-central-agent" containerID="cri-o://fbf261dc152da2105525608f99cd085cb605f4c4def074f0140516ec04df21eb" gracePeriod=30 Mar 21 05:09:49 crc kubenswrapper[4775]: I0321 05:09:49.674336 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerName="sg-core" containerID="cri-o://8f1be7b1e4c2dbdc319160849ad419dc210086da61fe3aad3c1277cd3f4787db" gracePeriod=30 Mar 21 05:09:49 crc kubenswrapper[4775]: I0321 05:09:49.674376 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerName="ceilometer-notification-agent" containerID="cri-o://e1897e25daedbe1665e342773e7750ff60df7bc2b0e9065d7163a3b3ead5e27d" gracePeriod=30 Mar 21 05:09:49 crc kubenswrapper[4775]: I0321 05:09:49.674403 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerName="proxy-httpd" containerID="cri-o://1019edd2ab7e9cd0ed64f98b04c2e3b4282ef0e3d503d9b2e72d9464a7033ab9" gracePeriod=30 Mar 21 05:09:49 crc kubenswrapper[4775]: I0321 05:09:49.698584 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.194:3000/\": EOF" Mar 21 05:09:49 crc kubenswrapper[4775]: I0321 05:09:49.860971 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.026143 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.205455 4775 generic.go:334] "Generic (PLEG): container finished" podID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerID="1019edd2ab7e9cd0ed64f98b04c2e3b4282ef0e3d503d9b2e72d9464a7033ab9" exitCode=0 Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.205488 4775 generic.go:334] "Generic (PLEG): container finished" podID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerID="8f1be7b1e4c2dbdc319160849ad419dc210086da61fe3aad3c1277cd3f4787db" exitCode=2 Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.205495 4775 generic.go:334] "Generic (PLEG): container finished" podID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerID="fbf261dc152da2105525608f99cd085cb605f4c4def074f0140516ec04df21eb" exitCode=0 Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.205546 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53","Type":"ContainerDied","Data":"1019edd2ab7e9cd0ed64f98b04c2e3b4282ef0e3d503d9b2e72d9464a7033ab9"} Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.205588 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53","Type":"ContainerDied","Data":"8f1be7b1e4c2dbdc319160849ad419dc210086da61fe3aad3c1277cd3f4787db"} Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.205642 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53","Type":"ContainerDied","Data":"fbf261dc152da2105525608f99cd085cb605f4c4def074f0140516ec04df21eb"} Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.207611 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" event={"ID":"d725aabc-ba32-4c0e-bc91-8819d73cae40","Type":"ContainerStarted","Data":"90f0ef44d50d3f68a6e80b10a816b133ae6b897ca8881964ca82e79aca5bf325"} Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.207723 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7d4895fa-5b3e-423f-85aa-030d2db076b3" containerName="nova-api-log" containerID="cri-o://ad57dd190fc32b9e6c21bd060d276edc5365c62c5bf0f9ffb91fab933879f448" gracePeriod=30 Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.207975 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7d4895fa-5b3e-423f-85aa-030d2db076b3" containerName="nova-api-api" containerID="cri-o://ec80243e9f9293a7a7355096362bcfe00d0fa0d7ef991f719431c39710253285" gracePeriod=30 Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.240686 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" podStartSLOduration=3.240671834 podStartE2EDuration="3.240671834s" podCreationTimestamp="2026-03-21 05:09:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:09:50.235471457 +0000 UTC m=+1343.211935081" watchObservedRunningTime="2026-03-21 05:09:50.240671834 +0000 UTC m=+1343.217135458" Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.725847 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.900980 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-combined-ca-bundle\") pod \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.901332 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-scripts\") pod \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.901460 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-log-httpd\") pod \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.901500 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-config-data\") pod \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.901525 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-run-httpd\") pod \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.901619 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsrdg\" (UniqueName: \"kubernetes.io/projected/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-kube-api-access-zsrdg\") pod \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.901639 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-sg-core-conf-yaml\") pod \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\" (UID: \"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53\") " Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.902271 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" (UID: "e3dd1c1b-7aba-4083-8a9c-037cd38b2d53"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.903326 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" (UID: "e3dd1c1b-7aba-4083-8a9c-037cd38b2d53"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.914222 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-scripts" (OuterVolumeSpecName: "scripts") pod "e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" (UID: "e3dd1c1b-7aba-4083-8a9c-037cd38b2d53"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.915193 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-kube-api-access-zsrdg" (OuterVolumeSpecName: "kube-api-access-zsrdg") pod "e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" (UID: "e3dd1c1b-7aba-4083-8a9c-037cd38b2d53"). InnerVolumeSpecName "kube-api-access-zsrdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.931239 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" (UID: "e3dd1c1b-7aba-4083-8a9c-037cd38b2d53"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:50 crc kubenswrapper[4775]: I0321 05:09:50.991547 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" (UID: "e3dd1c1b-7aba-4083-8a9c-037cd38b2d53"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.005884 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.005914 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.005924 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.005932 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.005940 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.005951 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsrdg\" (UniqueName: \"kubernetes.io/projected/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-kube-api-access-zsrdg\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.021491 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-config-data" (OuterVolumeSpecName: "config-data") pod "e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" (UID: "e3dd1c1b-7aba-4083-8a9c-037cd38b2d53"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.108242 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.218694 4775 generic.go:334] "Generic (PLEG): container finished" podID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerID="e1897e25daedbe1665e342773e7750ff60df7bc2b0e9065d7163a3b3ead5e27d" exitCode=0 Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.218739 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.218759 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53","Type":"ContainerDied","Data":"e1897e25daedbe1665e342773e7750ff60df7bc2b0e9065d7163a3b3ead5e27d"} Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.218978 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3dd1c1b-7aba-4083-8a9c-037cd38b2d53","Type":"ContainerDied","Data":"bd088ed9e7536178799aa0168c0a00723a8b5b8b146cbd4aa52d938131eb733b"} Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.219002 4775 scope.go:117] "RemoveContainer" containerID="1019edd2ab7e9cd0ed64f98b04c2e3b4282ef0e3d503d9b2e72d9464a7033ab9" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.221500 4775 generic.go:334] "Generic (PLEG): container finished" podID="7d4895fa-5b3e-423f-85aa-030d2db076b3" containerID="ad57dd190fc32b9e6c21bd060d276edc5365c62c5bf0f9ffb91fab933879f448" exitCode=143 Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.221582 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7d4895fa-5b3e-423f-85aa-030d2db076b3","Type":"ContainerDied","Data":"ad57dd190fc32b9e6c21bd060d276edc5365c62c5bf0f9ffb91fab933879f448"} Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.221638 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.249757 4775 scope.go:117] "RemoveContainer" containerID="8f1be7b1e4c2dbdc319160849ad419dc210086da61fe3aad3c1277cd3f4787db" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.254460 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.280705 4775 scope.go:117] "RemoveContainer" containerID="e1897e25daedbe1665e342773e7750ff60df7bc2b0e9065d7163a3b3ead5e27d" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.283488 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.294648 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:09:51 crc kubenswrapper[4775]: E0321 05:09:51.295206 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerName="proxy-httpd" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.295291 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerName="proxy-httpd" Mar 21 05:09:51 crc kubenswrapper[4775]: E0321 05:09:51.295319 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerName="ceilometer-notification-agent" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.295327 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerName="ceilometer-notification-agent" Mar 21 05:09:51 crc kubenswrapper[4775]: E0321 05:09:51.295350 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerName="ceilometer-central-agent" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.295359 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerName="ceilometer-central-agent" Mar 21 05:09:51 crc kubenswrapper[4775]: E0321 05:09:51.295376 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerName="sg-core" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.295385 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerName="sg-core" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.295670 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerName="ceilometer-notification-agent" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.295706 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerName="proxy-httpd" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.295752 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerName="ceilometer-central-agent" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.295770 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerName="sg-core" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.297997 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.305292 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.307482 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.307790 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.332065 4775 scope.go:117] "RemoveContainer" containerID="fbf261dc152da2105525608f99cd085cb605f4c4def074f0140516ec04df21eb" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.361983 4775 scope.go:117] "RemoveContainer" containerID="1019edd2ab7e9cd0ed64f98b04c2e3b4282ef0e3d503d9b2e72d9464a7033ab9" Mar 21 05:09:51 crc kubenswrapper[4775]: E0321 05:09:51.362595 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1019edd2ab7e9cd0ed64f98b04c2e3b4282ef0e3d503d9b2e72d9464a7033ab9\": container with ID starting with 1019edd2ab7e9cd0ed64f98b04c2e3b4282ef0e3d503d9b2e72d9464a7033ab9 not found: ID does not exist" containerID="1019edd2ab7e9cd0ed64f98b04c2e3b4282ef0e3d503d9b2e72d9464a7033ab9" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.362788 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1019edd2ab7e9cd0ed64f98b04c2e3b4282ef0e3d503d9b2e72d9464a7033ab9"} err="failed to get container status \"1019edd2ab7e9cd0ed64f98b04c2e3b4282ef0e3d503d9b2e72d9464a7033ab9\": rpc error: code = NotFound desc = could not find container \"1019edd2ab7e9cd0ed64f98b04c2e3b4282ef0e3d503d9b2e72d9464a7033ab9\": container with ID starting with 1019edd2ab7e9cd0ed64f98b04c2e3b4282ef0e3d503d9b2e72d9464a7033ab9 not found: ID does not exist" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.362893 4775 scope.go:117] "RemoveContainer" containerID="8f1be7b1e4c2dbdc319160849ad419dc210086da61fe3aad3c1277cd3f4787db" Mar 21 05:09:51 crc kubenswrapper[4775]: E0321 05:09:51.363313 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f1be7b1e4c2dbdc319160849ad419dc210086da61fe3aad3c1277cd3f4787db\": container with ID starting with 8f1be7b1e4c2dbdc319160849ad419dc210086da61fe3aad3c1277cd3f4787db not found: ID does not exist" containerID="8f1be7b1e4c2dbdc319160849ad419dc210086da61fe3aad3c1277cd3f4787db" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.363364 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f1be7b1e4c2dbdc319160849ad419dc210086da61fe3aad3c1277cd3f4787db"} err="failed to get container status \"8f1be7b1e4c2dbdc319160849ad419dc210086da61fe3aad3c1277cd3f4787db\": rpc error: code = NotFound desc = could not find container \"8f1be7b1e4c2dbdc319160849ad419dc210086da61fe3aad3c1277cd3f4787db\": container with ID starting with 8f1be7b1e4c2dbdc319160849ad419dc210086da61fe3aad3c1277cd3f4787db not found: ID does not exist" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.363400 4775 scope.go:117] "RemoveContainer" containerID="e1897e25daedbe1665e342773e7750ff60df7bc2b0e9065d7163a3b3ead5e27d" Mar 21 05:09:51 crc kubenswrapper[4775]: E0321 05:09:51.363780 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1897e25daedbe1665e342773e7750ff60df7bc2b0e9065d7163a3b3ead5e27d\": container with ID starting with e1897e25daedbe1665e342773e7750ff60df7bc2b0e9065d7163a3b3ead5e27d not found: ID does not exist" containerID="e1897e25daedbe1665e342773e7750ff60df7bc2b0e9065d7163a3b3ead5e27d" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.363813 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1897e25daedbe1665e342773e7750ff60df7bc2b0e9065d7163a3b3ead5e27d"} err="failed to get container status \"e1897e25daedbe1665e342773e7750ff60df7bc2b0e9065d7163a3b3ead5e27d\": rpc error: code = NotFound desc = could not find container \"e1897e25daedbe1665e342773e7750ff60df7bc2b0e9065d7163a3b3ead5e27d\": container with ID starting with e1897e25daedbe1665e342773e7750ff60df7bc2b0e9065d7163a3b3ead5e27d not found: ID does not exist" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.363832 4775 scope.go:117] "RemoveContainer" containerID="fbf261dc152da2105525608f99cd085cb605f4c4def074f0140516ec04df21eb" Mar 21 05:09:51 crc kubenswrapper[4775]: E0321 05:09:51.364076 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbf261dc152da2105525608f99cd085cb605f4c4def074f0140516ec04df21eb\": container with ID starting with fbf261dc152da2105525608f99cd085cb605f4c4def074f0140516ec04df21eb not found: ID does not exist" containerID="fbf261dc152da2105525608f99cd085cb605f4c4def074f0140516ec04df21eb" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.364205 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbf261dc152da2105525608f99cd085cb605f4c4def074f0140516ec04df21eb"} err="failed to get container status \"fbf261dc152da2105525608f99cd085cb605f4c4def074f0140516ec04df21eb\": rpc error: code = NotFound desc = could not find container \"fbf261dc152da2105525608f99cd085cb605f4c4def074f0140516ec04df21eb\": container with ID starting with fbf261dc152da2105525608f99cd085cb605f4c4def074f0140516ec04df21eb not found: ID does not exist" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.412906 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-log-httpd\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.412961 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d9sz\" (UniqueName: \"kubernetes.io/projected/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-kube-api-access-2d9sz\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.413005 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.413032 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-config-data\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.413071 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.413144 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-run-httpd\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.413174 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-scripts\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.514730 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-log-httpd\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.515075 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d9sz\" (UniqueName: \"kubernetes.io/projected/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-kube-api-access-2d9sz\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.515147 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.515177 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-config-data\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.515214 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.515265 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-run-httpd\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.515293 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-scripts\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.515902 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-log-httpd\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.516478 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-run-httpd\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.520740 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.521228 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.521737 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-config-data\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.526186 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-scripts\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.533275 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d9sz\" (UniqueName: \"kubernetes.io/projected/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-kube-api-access-2d9sz\") pod \"ceilometer-0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.628857 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.675265 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" path="/var/lib/kubelet/pods/e3dd1c1b-7aba-4083-8a9c-037cd38b2d53/volumes" Mar 21 05:09:51 crc kubenswrapper[4775]: I0321 05:09:51.724414 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:09:52 crc kubenswrapper[4775]: I0321 05:09:52.076236 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:09:52 crc kubenswrapper[4775]: W0321 05:09:52.078476 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e2c9d32_ac8d_41ed_ba32_d88dcaffe4b0.slice/crio-768b8ecf031f12d20a791f4f289282980e217faa8048c9e8011e33181c86d226 WatchSource:0}: Error finding container 768b8ecf031f12d20a791f4f289282980e217faa8048c9e8011e33181c86d226: Status 404 returned error can't find the container with id 768b8ecf031f12d20a791f4f289282980e217faa8048c9e8011e33181c86d226 Mar 21 05:09:52 crc kubenswrapper[4775]: I0321 05:09:52.230352 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0","Type":"ContainerStarted","Data":"768b8ecf031f12d20a791f4f289282980e217faa8048c9e8011e33181c86d226"} Mar 21 05:09:53 crc kubenswrapper[4775]: I0321 05:09:53.240209 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0","Type":"ContainerStarted","Data":"09a1cf40ceee0947e10538878f75aaf051ddd9cfb50d6c16bb5d54ed58dffeea"} Mar 21 05:09:53 crc kubenswrapper[4775]: I0321 05:09:53.799764 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 05:09:53 crc kubenswrapper[4775]: I0321 05:09:53.960464 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d4895fa-5b3e-423f-85aa-030d2db076b3-logs\") pod \"7d4895fa-5b3e-423f-85aa-030d2db076b3\" (UID: \"7d4895fa-5b3e-423f-85aa-030d2db076b3\") " Mar 21 05:09:53 crc kubenswrapper[4775]: I0321 05:09:53.960778 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d4895fa-5b3e-423f-85aa-030d2db076b3-combined-ca-bundle\") pod \"7d4895fa-5b3e-423f-85aa-030d2db076b3\" (UID: \"7d4895fa-5b3e-423f-85aa-030d2db076b3\") " Mar 21 05:09:53 crc kubenswrapper[4775]: I0321 05:09:53.960974 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krq52\" (UniqueName: \"kubernetes.io/projected/7d4895fa-5b3e-423f-85aa-030d2db076b3-kube-api-access-krq52\") pod \"7d4895fa-5b3e-423f-85aa-030d2db076b3\" (UID: \"7d4895fa-5b3e-423f-85aa-030d2db076b3\") " Mar 21 05:09:53 crc kubenswrapper[4775]: I0321 05:09:53.961089 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d4895fa-5b3e-423f-85aa-030d2db076b3-config-data\") pod \"7d4895fa-5b3e-423f-85aa-030d2db076b3\" (UID: \"7d4895fa-5b3e-423f-85aa-030d2db076b3\") " Mar 21 05:09:53 crc kubenswrapper[4775]: I0321 05:09:53.961261 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d4895fa-5b3e-423f-85aa-030d2db076b3-logs" (OuterVolumeSpecName: "logs") pod "7d4895fa-5b3e-423f-85aa-030d2db076b3" (UID: "7d4895fa-5b3e-423f-85aa-030d2db076b3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:09:53 crc kubenswrapper[4775]: I0321 05:09:53.961567 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d4895fa-5b3e-423f-85aa-030d2db076b3-logs\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:53 crc kubenswrapper[4775]: I0321 05:09:53.969231 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d4895fa-5b3e-423f-85aa-030d2db076b3-kube-api-access-krq52" (OuterVolumeSpecName: "kube-api-access-krq52") pod "7d4895fa-5b3e-423f-85aa-030d2db076b3" (UID: "7d4895fa-5b3e-423f-85aa-030d2db076b3"). InnerVolumeSpecName "kube-api-access-krq52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:09:53 crc kubenswrapper[4775]: E0321 05:09:53.997441 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d4895fa-5b3e-423f-85aa-030d2db076b3-config-data podName:7d4895fa-5b3e-423f-85aa-030d2db076b3 nodeName:}" failed. No retries permitted until 2026-03-21 05:09:54.497406806 +0000 UTC m=+1347.473870440 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/7d4895fa-5b3e-423f-85aa-030d2db076b3-config-data") pod "7d4895fa-5b3e-423f-85aa-030d2db076b3" (UID: "7d4895fa-5b3e-423f-85aa-030d2db076b3") : error deleting /var/lib/kubelet/pods/7d4895fa-5b3e-423f-85aa-030d2db076b3/volume-subpaths: remove /var/lib/kubelet/pods/7d4895fa-5b3e-423f-85aa-030d2db076b3/volume-subpaths: no such file or directory Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.001654 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d4895fa-5b3e-423f-85aa-030d2db076b3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d4895fa-5b3e-423f-85aa-030d2db076b3" (UID: "7d4895fa-5b3e-423f-85aa-030d2db076b3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.062902 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krq52\" (UniqueName: \"kubernetes.io/projected/7d4895fa-5b3e-423f-85aa-030d2db076b3-kube-api-access-krq52\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.062935 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d4895fa-5b3e-423f-85aa-030d2db076b3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.250993 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0","Type":"ContainerStarted","Data":"1d04a9e67e080d04335a90f9088873f5ae0ae56920f7eff7ab1e8e3ecca11bd2"} Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.253351 4775 generic.go:334] "Generic (PLEG): container finished" podID="7d4895fa-5b3e-423f-85aa-030d2db076b3" containerID="ec80243e9f9293a7a7355096362bcfe00d0fa0d7ef991f719431c39710253285" exitCode=0 Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.253378 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7d4895fa-5b3e-423f-85aa-030d2db076b3","Type":"ContainerDied","Data":"ec80243e9f9293a7a7355096362bcfe00d0fa0d7ef991f719431c39710253285"} Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.253395 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7d4895fa-5b3e-423f-85aa-030d2db076b3","Type":"ContainerDied","Data":"d95657026d4984394319083a88a788fdbce021127cb6cbf909a8f8735dc34db8"} Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.253411 4775 scope.go:117] "RemoveContainer" containerID="ec80243e9f9293a7a7355096362bcfe00d0fa0d7ef991f719431c39710253285" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.253456 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.273696 4775 scope.go:117] "RemoveContainer" containerID="ad57dd190fc32b9e6c21bd060d276edc5365c62c5bf0f9ffb91fab933879f448" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.363015 4775 scope.go:117] "RemoveContainer" containerID="ec80243e9f9293a7a7355096362bcfe00d0fa0d7ef991f719431c39710253285" Mar 21 05:09:54 crc kubenswrapper[4775]: E0321 05:09:54.363389 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec80243e9f9293a7a7355096362bcfe00d0fa0d7ef991f719431c39710253285\": container with ID starting with ec80243e9f9293a7a7355096362bcfe00d0fa0d7ef991f719431c39710253285 not found: ID does not exist" containerID="ec80243e9f9293a7a7355096362bcfe00d0fa0d7ef991f719431c39710253285" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.363424 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec80243e9f9293a7a7355096362bcfe00d0fa0d7ef991f719431c39710253285"} err="failed to get container status \"ec80243e9f9293a7a7355096362bcfe00d0fa0d7ef991f719431c39710253285\": rpc error: code = NotFound desc = could not find container \"ec80243e9f9293a7a7355096362bcfe00d0fa0d7ef991f719431c39710253285\": container with ID starting with ec80243e9f9293a7a7355096362bcfe00d0fa0d7ef991f719431c39710253285 not found: ID does not exist" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.363445 4775 scope.go:117] "RemoveContainer" containerID="ad57dd190fc32b9e6c21bd060d276edc5365c62c5bf0f9ffb91fab933879f448" Mar 21 05:09:54 crc kubenswrapper[4775]: E0321 05:09:54.363778 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad57dd190fc32b9e6c21bd060d276edc5365c62c5bf0f9ffb91fab933879f448\": container with ID starting with ad57dd190fc32b9e6c21bd060d276edc5365c62c5bf0f9ffb91fab933879f448 not found: ID does not exist" containerID="ad57dd190fc32b9e6c21bd060d276edc5365c62c5bf0f9ffb91fab933879f448" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.363836 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad57dd190fc32b9e6c21bd060d276edc5365c62c5bf0f9ffb91fab933879f448"} err="failed to get container status \"ad57dd190fc32b9e6c21bd060d276edc5365c62c5bf0f9ffb91fab933879f448\": rpc error: code = NotFound desc = could not find container \"ad57dd190fc32b9e6c21bd060d276edc5365c62c5bf0f9ffb91fab933879f448\": container with ID starting with ad57dd190fc32b9e6c21bd060d276edc5365c62c5bf0f9ffb91fab933879f448 not found: ID does not exist" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.573414 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d4895fa-5b3e-423f-85aa-030d2db076b3-config-data\") pod \"7d4895fa-5b3e-423f-85aa-030d2db076b3\" (UID: \"7d4895fa-5b3e-423f-85aa-030d2db076b3\") " Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.578270 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d4895fa-5b3e-423f-85aa-030d2db076b3-config-data" (OuterVolumeSpecName: "config-data") pod "7d4895fa-5b3e-423f-85aa-030d2db076b3" (UID: "7d4895fa-5b3e-423f-85aa-030d2db076b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.676199 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d4895fa-5b3e-423f-85aa-030d2db076b3-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.860837 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.890431 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.898734 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.911836 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.914806 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 21 05:09:54 crc kubenswrapper[4775]: E0321 05:09:54.915260 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d4895fa-5b3e-423f-85aa-030d2db076b3" containerName="nova-api-log" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.915282 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d4895fa-5b3e-423f-85aa-030d2db076b3" containerName="nova-api-log" Mar 21 05:09:54 crc kubenswrapper[4775]: E0321 05:09:54.915310 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d4895fa-5b3e-423f-85aa-030d2db076b3" containerName="nova-api-api" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.915318 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d4895fa-5b3e-423f-85aa-030d2db076b3" containerName="nova-api-api" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.915495 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d4895fa-5b3e-423f-85aa-030d2db076b3" containerName="nova-api-api" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.915516 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d4895fa-5b3e-423f-85aa-030d2db076b3" containerName="nova-api-log" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.916478 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.923617 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.925530 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.925568 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 21 05:09:54 crc kubenswrapper[4775]: I0321 05:09:54.931654 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.082740 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jdpm\" (UniqueName: \"kubernetes.io/projected/fb3f1e05-6056-4a7f-9452-22caa26e74fd-kube-api-access-8jdpm\") pod \"nova-api-0\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " pod="openstack/nova-api-0" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.082839 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-config-data\") pod \"nova-api-0\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " pod="openstack/nova-api-0" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.082866 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb3f1e05-6056-4a7f-9452-22caa26e74fd-logs\") pod \"nova-api-0\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " pod="openstack/nova-api-0" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.082886 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " pod="openstack/nova-api-0" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.082917 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-public-tls-certs\") pod \"nova-api-0\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " pod="openstack/nova-api-0" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.083310 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-internal-tls-certs\") pod \"nova-api-0\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " pod="openstack/nova-api-0" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.185174 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-config-data\") pod \"nova-api-0\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " pod="openstack/nova-api-0" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.185237 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb3f1e05-6056-4a7f-9452-22caa26e74fd-logs\") pod \"nova-api-0\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " pod="openstack/nova-api-0" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.185273 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " pod="openstack/nova-api-0" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.185315 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-public-tls-certs\") pod \"nova-api-0\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " pod="openstack/nova-api-0" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.185428 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-internal-tls-certs\") pod \"nova-api-0\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " pod="openstack/nova-api-0" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.185519 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jdpm\" (UniqueName: \"kubernetes.io/projected/fb3f1e05-6056-4a7f-9452-22caa26e74fd-kube-api-access-8jdpm\") pod \"nova-api-0\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " pod="openstack/nova-api-0" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.185756 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb3f1e05-6056-4a7f-9452-22caa26e74fd-logs\") pod \"nova-api-0\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " pod="openstack/nova-api-0" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.189531 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " pod="openstack/nova-api-0" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.191040 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-config-data\") pod \"nova-api-0\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " pod="openstack/nova-api-0" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.191058 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-internal-tls-certs\") pod \"nova-api-0\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " pod="openstack/nova-api-0" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.197584 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-public-tls-certs\") pod \"nova-api-0\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " pod="openstack/nova-api-0" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.203232 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jdpm\" (UniqueName: \"kubernetes.io/projected/fb3f1e05-6056-4a7f-9452-22caa26e74fd-kube-api-access-8jdpm\") pod \"nova-api-0\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " pod="openstack/nova-api-0" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.232564 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.268138 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0","Type":"ContainerStarted","Data":"ce4a1712fe9b7303080889248fbdbd025014f38760cd0cb19ad06dd199164a50"} Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.290375 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.485221 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-s5ct6"] Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.486976 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-s5ct6" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.489287 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.490596 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.497158 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-s5ct6"] Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.593235 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnjnj\" (UniqueName: \"kubernetes.io/projected/f3172d76-da59-4dee-95df-6b74b7fb0033-kube-api-access-fnjnj\") pod \"nova-cell1-cell-mapping-s5ct6\" (UID: \"f3172d76-da59-4dee-95df-6b74b7fb0033\") " pod="openstack/nova-cell1-cell-mapping-s5ct6" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.593433 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3172d76-da59-4dee-95df-6b74b7fb0033-scripts\") pod \"nova-cell1-cell-mapping-s5ct6\" (UID: \"f3172d76-da59-4dee-95df-6b74b7fb0033\") " pod="openstack/nova-cell1-cell-mapping-s5ct6" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.593508 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3172d76-da59-4dee-95df-6b74b7fb0033-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-s5ct6\" (UID: \"f3172d76-da59-4dee-95df-6b74b7fb0033\") " pod="openstack/nova-cell1-cell-mapping-s5ct6" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.593602 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3172d76-da59-4dee-95df-6b74b7fb0033-config-data\") pod \"nova-cell1-cell-mapping-s5ct6\" (UID: \"f3172d76-da59-4dee-95df-6b74b7fb0033\") " pod="openstack/nova-cell1-cell-mapping-s5ct6" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.674956 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d4895fa-5b3e-423f-85aa-030d2db076b3" path="/var/lib/kubelet/pods/7d4895fa-5b3e-423f-85aa-030d2db076b3/volumes" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.695447 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnjnj\" (UniqueName: \"kubernetes.io/projected/f3172d76-da59-4dee-95df-6b74b7fb0033-kube-api-access-fnjnj\") pod \"nova-cell1-cell-mapping-s5ct6\" (UID: \"f3172d76-da59-4dee-95df-6b74b7fb0033\") " pod="openstack/nova-cell1-cell-mapping-s5ct6" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.695585 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3172d76-da59-4dee-95df-6b74b7fb0033-scripts\") pod \"nova-cell1-cell-mapping-s5ct6\" (UID: \"f3172d76-da59-4dee-95df-6b74b7fb0033\") " pod="openstack/nova-cell1-cell-mapping-s5ct6" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.695625 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3172d76-da59-4dee-95df-6b74b7fb0033-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-s5ct6\" (UID: \"f3172d76-da59-4dee-95df-6b74b7fb0033\") " pod="openstack/nova-cell1-cell-mapping-s5ct6" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.695668 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3172d76-da59-4dee-95df-6b74b7fb0033-config-data\") pod \"nova-cell1-cell-mapping-s5ct6\" (UID: \"f3172d76-da59-4dee-95df-6b74b7fb0033\") " pod="openstack/nova-cell1-cell-mapping-s5ct6" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.701372 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3172d76-da59-4dee-95df-6b74b7fb0033-config-data\") pod \"nova-cell1-cell-mapping-s5ct6\" (UID: \"f3172d76-da59-4dee-95df-6b74b7fb0033\") " pod="openstack/nova-cell1-cell-mapping-s5ct6" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.703676 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3172d76-da59-4dee-95df-6b74b7fb0033-scripts\") pod \"nova-cell1-cell-mapping-s5ct6\" (UID: \"f3172d76-da59-4dee-95df-6b74b7fb0033\") " pod="openstack/nova-cell1-cell-mapping-s5ct6" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.704029 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3172d76-da59-4dee-95df-6b74b7fb0033-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-s5ct6\" (UID: \"f3172d76-da59-4dee-95df-6b74b7fb0033\") " pod="openstack/nova-cell1-cell-mapping-s5ct6" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.713825 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.725090 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnjnj\" (UniqueName: \"kubernetes.io/projected/f3172d76-da59-4dee-95df-6b74b7fb0033-kube-api-access-fnjnj\") pod \"nova-cell1-cell-mapping-s5ct6\" (UID: \"f3172d76-da59-4dee-95df-6b74b7fb0033\") " pod="openstack/nova-cell1-cell-mapping-s5ct6" Mar 21 05:09:55 crc kubenswrapper[4775]: I0321 05:09:55.817815 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-s5ct6" Mar 21 05:09:56 crc kubenswrapper[4775]: I0321 05:09:56.275559 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-s5ct6"] Mar 21 05:09:56 crc kubenswrapper[4775]: I0321 05:09:56.292551 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fb3f1e05-6056-4a7f-9452-22caa26e74fd","Type":"ContainerStarted","Data":"4fa32c27636fc153967f4aaaf1df73d8fca4656fe14bc2f63c7966be0877d999"} Mar 21 05:09:56 crc kubenswrapper[4775]: I0321 05:09:56.292604 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fb3f1e05-6056-4a7f-9452-22caa26e74fd","Type":"ContainerStarted","Data":"5417129905c56019139e8fe84dd07140ab4c584dde31a376f01ba295d9052432"} Mar 21 05:09:56 crc kubenswrapper[4775]: I0321 05:09:56.292627 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fb3f1e05-6056-4a7f-9452-22caa26e74fd","Type":"ContainerStarted","Data":"58500c9c4082d59c6b4aa8467cacee4491bef38fcd23aebf0ec570b68876648c"} Mar 21 05:09:56 crc kubenswrapper[4775]: I0321 05:09:56.297968 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0","Type":"ContainerStarted","Data":"c5a4cae04fe0bc9abff39e7ed6ec2b33edafb6e8a7275a05a5ea9574dcb70f60"} Mar 21 05:09:56 crc kubenswrapper[4775]: I0321 05:09:56.298134 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" containerName="ceilometer-central-agent" containerID="cri-o://09a1cf40ceee0947e10538878f75aaf051ddd9cfb50d6c16bb5d54ed58dffeea" gracePeriod=30 Mar 21 05:09:56 crc kubenswrapper[4775]: I0321 05:09:56.298334 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 21 05:09:56 crc kubenswrapper[4775]: I0321 05:09:56.298393 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" containerName="proxy-httpd" containerID="cri-o://c5a4cae04fe0bc9abff39e7ed6ec2b33edafb6e8a7275a05a5ea9574dcb70f60" gracePeriod=30 Mar 21 05:09:56 crc kubenswrapper[4775]: I0321 05:09:56.298456 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" containerName="sg-core" containerID="cri-o://ce4a1712fe9b7303080889248fbdbd025014f38760cd0cb19ad06dd199164a50" gracePeriod=30 Mar 21 05:09:56 crc kubenswrapper[4775]: I0321 05:09:56.298503 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" containerName="ceilometer-notification-agent" containerID="cri-o://1d04a9e67e080d04335a90f9088873f5ae0ae56920f7eff7ab1e8e3ecca11bd2" gracePeriod=30 Mar 21 05:09:56 crc kubenswrapper[4775]: I0321 05:09:56.311866 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-s5ct6" event={"ID":"f3172d76-da59-4dee-95df-6b74b7fb0033","Type":"ContainerStarted","Data":"d597814d14b2881f772c64b63db812d2cc5083310896232b321ea29b19b85684"} Mar 21 05:09:56 crc kubenswrapper[4775]: I0321 05:09:56.315639 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.315617631 podStartE2EDuration="2.315617631s" podCreationTimestamp="2026-03-21 05:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:09:56.312289017 +0000 UTC m=+1349.288752651" watchObservedRunningTime="2026-03-21 05:09:56.315617631 +0000 UTC m=+1349.292081255" Mar 21 05:09:56 crc kubenswrapper[4775]: I0321 05:09:56.335615 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.460152588 podStartE2EDuration="5.335591336s" podCreationTimestamp="2026-03-21 05:09:51 +0000 UTC" firstStartedPulling="2026-03-21 05:09:52.080920327 +0000 UTC m=+1345.057383941" lastFinishedPulling="2026-03-21 05:09:55.956359065 +0000 UTC m=+1348.932822689" observedRunningTime="2026-03-21 05:09:56.329527024 +0000 UTC m=+1349.305990648" watchObservedRunningTime="2026-03-21 05:09:56.335591336 +0000 UTC m=+1349.312054970" Mar 21 05:09:57 crc kubenswrapper[4775]: I0321 05:09:57.339681 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-s5ct6" event={"ID":"f3172d76-da59-4dee-95df-6b74b7fb0033","Type":"ContainerStarted","Data":"244997576ccb04a6089eebd186997e7f409b2534428a82f578796006c7c674ed"} Mar 21 05:09:57 crc kubenswrapper[4775]: I0321 05:09:57.345312 4775 generic.go:334] "Generic (PLEG): container finished" podID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" containerID="c5a4cae04fe0bc9abff39e7ed6ec2b33edafb6e8a7275a05a5ea9574dcb70f60" exitCode=0 Mar 21 05:09:57 crc kubenswrapper[4775]: I0321 05:09:57.345341 4775 generic.go:334] "Generic (PLEG): container finished" podID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" containerID="ce4a1712fe9b7303080889248fbdbd025014f38760cd0cb19ad06dd199164a50" exitCode=2 Mar 21 05:09:57 crc kubenswrapper[4775]: I0321 05:09:57.345350 4775 generic.go:334] "Generic (PLEG): container finished" podID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" containerID="1d04a9e67e080d04335a90f9088873f5ae0ae56920f7eff7ab1e8e3ecca11bd2" exitCode=0 Mar 21 05:09:57 crc kubenswrapper[4775]: I0321 05:09:57.346111 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0","Type":"ContainerDied","Data":"c5a4cae04fe0bc9abff39e7ed6ec2b33edafb6e8a7275a05a5ea9574dcb70f60"} Mar 21 05:09:57 crc kubenswrapper[4775]: I0321 05:09:57.346160 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0","Type":"ContainerDied","Data":"ce4a1712fe9b7303080889248fbdbd025014f38760cd0cb19ad06dd199164a50"} Mar 21 05:09:57 crc kubenswrapper[4775]: I0321 05:09:57.346173 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0","Type":"ContainerDied","Data":"1d04a9e67e080d04335a90f9088873f5ae0ae56920f7eff7ab1e8e3ecca11bd2"} Mar 21 05:09:57 crc kubenswrapper[4775]: I0321 05:09:57.368784 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-s5ct6" podStartSLOduration=2.368747243 podStartE2EDuration="2.368747243s" podCreationTimestamp="2026-03-21 05:09:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:09:57.360533311 +0000 UTC m=+1350.336996935" watchObservedRunningTime="2026-03-21 05:09:57.368747243 +0000 UTC m=+1350.345210867" Mar 21 05:09:57 crc kubenswrapper[4775]: I0321 05:09:57.742280 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:09:57 crc kubenswrapper[4775]: I0321 05:09:57.809470 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-qsbbk"] Mar 21 05:09:57 crc kubenswrapper[4775]: I0321 05:09:57.809754 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" podUID="20a97ec2-24be-494f-a3e8-dc7d202021c3" containerName="dnsmasq-dns" containerID="cri-o://210c52f72b30e6d06b1636b7f4ba7c385a206b219bca52c176d9da81b984a93a" gracePeriod=10 Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.356870 4775 generic.go:334] "Generic (PLEG): container finished" podID="20a97ec2-24be-494f-a3e8-dc7d202021c3" containerID="210c52f72b30e6d06b1636b7f4ba7c385a206b219bca52c176d9da81b984a93a" exitCode=0 Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.357582 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" event={"ID":"20a97ec2-24be-494f-a3e8-dc7d202021c3","Type":"ContainerDied","Data":"210c52f72b30e6d06b1636b7f4ba7c385a206b219bca52c176d9da81b984a93a"} Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.357610 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" event={"ID":"20a97ec2-24be-494f-a3e8-dc7d202021c3","Type":"ContainerDied","Data":"9a2e5054b120d9d9b114985bfe9a3a6397285644c20c24699a021ae58fcb7c61"} Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.357620 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a2e5054b120d9d9b114985bfe9a3a6397285644c20c24699a021ae58fcb7c61" Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.410503 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.455374 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-dns-swift-storage-0\") pod \"20a97ec2-24be-494f-a3e8-dc7d202021c3\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.455506 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-dns-svc\") pod \"20a97ec2-24be-494f-a3e8-dc7d202021c3\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.455552 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-ovsdbserver-sb\") pod \"20a97ec2-24be-494f-a3e8-dc7d202021c3\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.455580 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptvg9\" (UniqueName: \"kubernetes.io/projected/20a97ec2-24be-494f-a3e8-dc7d202021c3-kube-api-access-ptvg9\") pod \"20a97ec2-24be-494f-a3e8-dc7d202021c3\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.455690 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-config\") pod \"20a97ec2-24be-494f-a3e8-dc7d202021c3\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.455727 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-ovsdbserver-nb\") pod \"20a97ec2-24be-494f-a3e8-dc7d202021c3\" (UID: \"20a97ec2-24be-494f-a3e8-dc7d202021c3\") " Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.472347 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20a97ec2-24be-494f-a3e8-dc7d202021c3-kube-api-access-ptvg9" (OuterVolumeSpecName: "kube-api-access-ptvg9") pod "20a97ec2-24be-494f-a3e8-dc7d202021c3" (UID: "20a97ec2-24be-494f-a3e8-dc7d202021c3"). InnerVolumeSpecName "kube-api-access-ptvg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.529643 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "20a97ec2-24be-494f-a3e8-dc7d202021c3" (UID: "20a97ec2-24be-494f-a3e8-dc7d202021c3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.531774 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "20a97ec2-24be-494f-a3e8-dc7d202021c3" (UID: "20a97ec2-24be-494f-a3e8-dc7d202021c3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.536050 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "20a97ec2-24be-494f-a3e8-dc7d202021c3" (UID: "20a97ec2-24be-494f-a3e8-dc7d202021c3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.557863 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "20a97ec2-24be-494f-a3e8-dc7d202021c3" (UID: "20a97ec2-24be-494f-a3e8-dc7d202021c3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.559535 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.559567 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.559581 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptvg9\" (UniqueName: \"kubernetes.io/projected/20a97ec2-24be-494f-a3e8-dc7d202021c3-kube-api-access-ptvg9\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.559590 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.559601 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.570764 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-config" (OuterVolumeSpecName: "config") pod "20a97ec2-24be-494f-a3e8-dc7d202021c3" (UID: "20a97ec2-24be-494f-a3e8-dc7d202021c3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:09:58 crc kubenswrapper[4775]: I0321 05:09:58.662391 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20a97ec2-24be-494f-a3e8-dc7d202021c3-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.371530 4775 generic.go:334] "Generic (PLEG): container finished" podID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" containerID="09a1cf40ceee0947e10538878f75aaf051ddd9cfb50d6c16bb5d54ed58dffeea" exitCode=0 Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.371631 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0","Type":"ContainerDied","Data":"09a1cf40ceee0947e10538878f75aaf051ddd9cfb50d6c16bb5d54ed58dffeea"} Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.371945 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.430744 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-qsbbk"] Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.476089 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-qsbbk"] Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.673347 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20a97ec2-24be-494f-a3e8-dc7d202021c3" path="/var/lib/kubelet/pods/20a97ec2-24be-494f-a3e8-dc7d202021c3/volumes" Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.731594 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.784823 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-log-httpd\") pod \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.784989 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-sg-core-conf-yaml\") pod \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.785058 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d9sz\" (UniqueName: \"kubernetes.io/projected/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-kube-api-access-2d9sz\") pod \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.785137 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-combined-ca-bundle\") pod \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.785229 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-run-httpd\") pod \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.785280 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-config-data\") pod \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.785296 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" (UID: "3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.785324 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-scripts\") pod \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\" (UID: \"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0\") " Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.785738 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" (UID: "3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.786029 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.786053 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.794279 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-scripts" (OuterVolumeSpecName: "scripts") pod "3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" (UID: "3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.794447 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-kube-api-access-2d9sz" (OuterVolumeSpecName: "kube-api-access-2d9sz") pod "3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" (UID: "3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0"). InnerVolumeSpecName "kube-api-access-2d9sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.813844 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" (UID: "3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.880998 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" (UID: "3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.888322 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d9sz\" (UniqueName: \"kubernetes.io/projected/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-kube-api-access-2d9sz\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.888367 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.888377 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.888385 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.899418 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-config-data" (OuterVolumeSpecName: "config-data") pod "3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" (UID: "3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:09:59 crc kubenswrapper[4775]: I0321 05:09:59.989866 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.143927 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567830-jv78x"] Mar 21 05:10:00 crc kubenswrapper[4775]: E0321 05:10:00.144383 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" containerName="sg-core" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.144407 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" containerName="sg-core" Mar 21 05:10:00 crc kubenswrapper[4775]: E0321 05:10:00.144430 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a97ec2-24be-494f-a3e8-dc7d202021c3" containerName="dnsmasq-dns" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.144438 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a97ec2-24be-494f-a3e8-dc7d202021c3" containerName="dnsmasq-dns" Mar 21 05:10:00 crc kubenswrapper[4775]: E0321 05:10:00.144461 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" containerName="ceilometer-central-agent" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.144469 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" containerName="ceilometer-central-agent" Mar 21 05:10:00 crc kubenswrapper[4775]: E0321 05:10:00.144488 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" containerName="ceilometer-notification-agent" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.144496 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" containerName="ceilometer-notification-agent" Mar 21 05:10:00 crc kubenswrapper[4775]: E0321 05:10:00.144510 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a97ec2-24be-494f-a3e8-dc7d202021c3" containerName="init" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.144517 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a97ec2-24be-494f-a3e8-dc7d202021c3" containerName="init" Mar 21 05:10:00 crc kubenswrapper[4775]: E0321 05:10:00.144532 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" containerName="proxy-httpd" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.144540 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" containerName="proxy-httpd" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.144735 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" containerName="proxy-httpd" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.144759 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" containerName="ceilometer-central-agent" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.144774 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="20a97ec2-24be-494f-a3e8-dc7d202021c3" containerName="dnsmasq-dns" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.144785 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" containerName="sg-core" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.144803 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" containerName="ceilometer-notification-agent" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.145528 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567830-jv78x" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.149657 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.149689 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.149917 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.155547 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567830-jv78x"] Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.193452 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns84k\" (UniqueName: \"kubernetes.io/projected/33cb4f91-aca0-4221-873f-e2c16e30ccee-kube-api-access-ns84k\") pod \"auto-csr-approver-29567830-jv78x\" (UID: \"33cb4f91-aca0-4221-873f-e2c16e30ccee\") " pod="openshift-infra/auto-csr-approver-29567830-jv78x" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.295199 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns84k\" (UniqueName: \"kubernetes.io/projected/33cb4f91-aca0-4221-873f-e2c16e30ccee-kube-api-access-ns84k\") pod \"auto-csr-approver-29567830-jv78x\" (UID: \"33cb4f91-aca0-4221-873f-e2c16e30ccee\") " pod="openshift-infra/auto-csr-approver-29567830-jv78x" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.314817 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns84k\" (UniqueName: \"kubernetes.io/projected/33cb4f91-aca0-4221-873f-e2c16e30ccee-kube-api-access-ns84k\") pod \"auto-csr-approver-29567830-jv78x\" (UID: \"33cb4f91-aca0-4221-873f-e2c16e30ccee\") " pod="openshift-infra/auto-csr-approver-29567830-jv78x" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.395858 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0","Type":"ContainerDied","Data":"768b8ecf031f12d20a791f4f289282980e217faa8048c9e8011e33181c86d226"} Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.396973 4775 scope.go:117] "RemoveContainer" containerID="c5a4cae04fe0bc9abff39e7ed6ec2b33edafb6e8a7275a05a5ea9574dcb70f60" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.395921 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.448330 4775 scope.go:117] "RemoveContainer" containerID="ce4a1712fe9b7303080889248fbdbd025014f38760cd0cb19ad06dd199164a50" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.454524 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.463536 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.465268 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567830-jv78x" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.479919 4775 scope.go:117] "RemoveContainer" containerID="1d04a9e67e080d04335a90f9088873f5ae0ae56920f7eff7ab1e8e3ecca11bd2" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.482430 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.485353 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.489100 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.489715 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.499827 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.528859 4775 scope.go:117] "RemoveContainer" containerID="09a1cf40ceee0947e10538878f75aaf051ddd9cfb50d6c16bb5d54ed58dffeea" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.600048 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk4fk\" (UniqueName: \"kubernetes.io/projected/7fdf78a6-b107-4949-affa-6152d15afda0-kube-api-access-lk4fk\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.600145 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.600184 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-config-data\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.600269 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fdf78a6-b107-4949-affa-6152d15afda0-log-httpd\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.600298 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fdf78a6-b107-4949-affa-6152d15afda0-run-httpd\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.600345 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.600436 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-scripts\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.701949 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.702456 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-scripts\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.702581 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk4fk\" (UniqueName: \"kubernetes.io/projected/7fdf78a6-b107-4949-affa-6152d15afda0-kube-api-access-lk4fk\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.702673 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.702706 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-config-data\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.702731 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fdf78a6-b107-4949-affa-6152d15afda0-log-httpd\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.702757 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fdf78a6-b107-4949-affa-6152d15afda0-run-httpd\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.704239 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fdf78a6-b107-4949-affa-6152d15afda0-log-httpd\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.704335 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fdf78a6-b107-4949-affa-6152d15afda0-run-httpd\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.706829 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-scripts\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.720110 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.720783 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.721829 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-config-data\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.725544 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk4fk\" (UniqueName: \"kubernetes.io/projected/7fdf78a6-b107-4949-affa-6152d15afda0-kube-api-access-lk4fk\") pod \"ceilometer-0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.849410 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:10:00 crc kubenswrapper[4775]: I0321 05:10:00.969424 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567830-jv78x"] Mar 21 05:10:00 crc kubenswrapper[4775]: W0321 05:10:00.983386 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33cb4f91_aca0_4221_873f_e2c16e30ccee.slice/crio-e13af9720ea7d69a98c04ade641d3957d885af03c65baffd94332be373fb4ff3 WatchSource:0}: Error finding container e13af9720ea7d69a98c04ade641d3957d885af03c65baffd94332be373fb4ff3: Status 404 returned error can't find the container with id e13af9720ea7d69a98c04ade641d3957d885af03c65baffd94332be373fb4ff3 Mar 21 05:10:01 crc kubenswrapper[4775]: I0321 05:10:01.324673 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:10:01 crc kubenswrapper[4775]: I0321 05:10:01.407646 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567830-jv78x" event={"ID":"33cb4f91-aca0-4221-873f-e2c16e30ccee","Type":"ContainerStarted","Data":"e13af9720ea7d69a98c04ade641d3957d885af03c65baffd94332be373fb4ff3"} Mar 21 05:10:01 crc kubenswrapper[4775]: I0321 05:10:01.411349 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fdf78a6-b107-4949-affa-6152d15afda0","Type":"ContainerStarted","Data":"0199a14e4f5a12a8019f727e78e7f15430a4ee84de969cda88096ca15bbed15b"} Mar 21 05:10:01 crc kubenswrapper[4775]: I0321 05:10:01.673489 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0" path="/var/lib/kubelet/pods/3e2c9d32-ac8d-41ed-ba32-d88dcaffe4b0/volumes" Mar 21 05:10:02 crc kubenswrapper[4775]: I0321 05:10:02.422189 4775 generic.go:334] "Generic (PLEG): container finished" podID="f3172d76-da59-4dee-95df-6b74b7fb0033" containerID="244997576ccb04a6089eebd186997e7f409b2534428a82f578796006c7c674ed" exitCode=0 Mar 21 05:10:02 crc kubenswrapper[4775]: I0321 05:10:02.422305 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-s5ct6" event={"ID":"f3172d76-da59-4dee-95df-6b74b7fb0033","Type":"ContainerDied","Data":"244997576ccb04a6089eebd186997e7f409b2534428a82f578796006c7c674ed"} Mar 21 05:10:02 crc kubenswrapper[4775]: I0321 05:10:02.482333 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:10:02 crc kubenswrapper[4775]: I0321 05:10:02.482412 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:10:03 crc kubenswrapper[4775]: I0321 05:10:03.106752 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-bccf8f775-qsbbk" podUID="20a97ec2-24be-494f-a3e8-dc7d202021c3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.190:5353: i/o timeout" Mar 21 05:10:03 crc kubenswrapper[4775]: I0321 05:10:03.432745 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fdf78a6-b107-4949-affa-6152d15afda0","Type":"ContainerStarted","Data":"8f8cabcb03903e4d89761a18703dd5c97870cd2e3d0ef375a4910ff30a8d99cf"} Mar 21 05:10:03 crc kubenswrapper[4775]: I0321 05:10:03.794422 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-s5ct6" Mar 21 05:10:03 crc kubenswrapper[4775]: I0321 05:10:03.862720 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3172d76-da59-4dee-95df-6b74b7fb0033-scripts\") pod \"f3172d76-da59-4dee-95df-6b74b7fb0033\" (UID: \"f3172d76-da59-4dee-95df-6b74b7fb0033\") " Mar 21 05:10:03 crc kubenswrapper[4775]: I0321 05:10:03.862813 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3172d76-da59-4dee-95df-6b74b7fb0033-config-data\") pod \"f3172d76-da59-4dee-95df-6b74b7fb0033\" (UID: \"f3172d76-da59-4dee-95df-6b74b7fb0033\") " Mar 21 05:10:03 crc kubenswrapper[4775]: I0321 05:10:03.862921 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnjnj\" (UniqueName: \"kubernetes.io/projected/f3172d76-da59-4dee-95df-6b74b7fb0033-kube-api-access-fnjnj\") pod \"f3172d76-da59-4dee-95df-6b74b7fb0033\" (UID: \"f3172d76-da59-4dee-95df-6b74b7fb0033\") " Mar 21 05:10:03 crc kubenswrapper[4775]: I0321 05:10:03.862978 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3172d76-da59-4dee-95df-6b74b7fb0033-combined-ca-bundle\") pod \"f3172d76-da59-4dee-95df-6b74b7fb0033\" (UID: \"f3172d76-da59-4dee-95df-6b74b7fb0033\") " Mar 21 05:10:03 crc kubenswrapper[4775]: I0321 05:10:03.868471 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3172d76-da59-4dee-95df-6b74b7fb0033-scripts" (OuterVolumeSpecName: "scripts") pod "f3172d76-da59-4dee-95df-6b74b7fb0033" (UID: "f3172d76-da59-4dee-95df-6b74b7fb0033"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:10:03 crc kubenswrapper[4775]: I0321 05:10:03.868945 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3172d76-da59-4dee-95df-6b74b7fb0033-kube-api-access-fnjnj" (OuterVolumeSpecName: "kube-api-access-fnjnj") pod "f3172d76-da59-4dee-95df-6b74b7fb0033" (UID: "f3172d76-da59-4dee-95df-6b74b7fb0033"). InnerVolumeSpecName "kube-api-access-fnjnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:10:03 crc kubenswrapper[4775]: I0321 05:10:03.894238 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3172d76-da59-4dee-95df-6b74b7fb0033-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f3172d76-da59-4dee-95df-6b74b7fb0033" (UID: "f3172d76-da59-4dee-95df-6b74b7fb0033"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:10:03 crc kubenswrapper[4775]: I0321 05:10:03.894608 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3172d76-da59-4dee-95df-6b74b7fb0033-config-data" (OuterVolumeSpecName: "config-data") pod "f3172d76-da59-4dee-95df-6b74b7fb0033" (UID: "f3172d76-da59-4dee-95df-6b74b7fb0033"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:10:03 crc kubenswrapper[4775]: I0321 05:10:03.964732 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3172d76-da59-4dee-95df-6b74b7fb0033-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:03 crc kubenswrapper[4775]: I0321 05:10:03.964761 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3172d76-da59-4dee-95df-6b74b7fb0033-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:03 crc kubenswrapper[4775]: I0321 05:10:03.964770 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3172d76-da59-4dee-95df-6b74b7fb0033-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:03 crc kubenswrapper[4775]: I0321 05:10:03.964778 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnjnj\" (UniqueName: \"kubernetes.io/projected/f3172d76-da59-4dee-95df-6b74b7fb0033-kube-api-access-fnjnj\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:04 crc kubenswrapper[4775]: I0321 05:10:04.444997 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-s5ct6" event={"ID":"f3172d76-da59-4dee-95df-6b74b7fb0033","Type":"ContainerDied","Data":"d597814d14b2881f772c64b63db812d2cc5083310896232b321ea29b19b85684"} Mar 21 05:10:04 crc kubenswrapper[4775]: I0321 05:10:04.445365 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d597814d14b2881f772c64b63db812d2cc5083310896232b321ea29b19b85684" Mar 21 05:10:04 crc kubenswrapper[4775]: I0321 05:10:04.445438 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-s5ct6" Mar 21 05:10:04 crc kubenswrapper[4775]: I0321 05:10:04.632573 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 05:10:04 crc kubenswrapper[4775]: I0321 05:10:04.632804 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="0f713501-fa26-4ef2-87ac-c5376b84891d" containerName="nova-scheduler-scheduler" containerID="cri-o://c561875841331ee73e6bcb2e365ce80407bc0539d3ec5ad570942023fedeb9c2" gracePeriod=30 Mar 21 05:10:04 crc kubenswrapper[4775]: I0321 05:10:04.650762 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 21 05:10:04 crc kubenswrapper[4775]: I0321 05:10:04.651079 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="fb3f1e05-6056-4a7f-9452-22caa26e74fd" containerName="nova-api-log" containerID="cri-o://5417129905c56019139e8fe84dd07140ab4c584dde31a376f01ba295d9052432" gracePeriod=30 Mar 21 05:10:04 crc kubenswrapper[4775]: I0321 05:10:04.651261 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="fb3f1e05-6056-4a7f-9452-22caa26e74fd" containerName="nova-api-api" containerID="cri-o://4fa32c27636fc153967f4aaaf1df73d8fca4656fe14bc2f63c7966be0877d999" gracePeriod=30 Mar 21 05:10:04 crc kubenswrapper[4775]: I0321 05:10:04.666606 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:10:04 crc kubenswrapper[4775]: I0321 05:10:04.666834 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="636228b7-669b-4b5d-abec-bf78cb1513f0" containerName="nova-metadata-log" containerID="cri-o://9a436c521b9cde86255d5d1bb86e7e7617657d67544c27a71666b10506ae93dc" gracePeriod=30 Mar 21 05:10:04 crc kubenswrapper[4775]: I0321 05:10:04.666964 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="636228b7-669b-4b5d-abec-bf78cb1513f0" containerName="nova-metadata-metadata" containerID="cri-o://88b942c0d478eda3ab139d99845879b76d209947d510208b64a88f8f909afefc" gracePeriod=30 Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.459584 4775 generic.go:334] "Generic (PLEG): container finished" podID="fb3f1e05-6056-4a7f-9452-22caa26e74fd" containerID="4fa32c27636fc153967f4aaaf1df73d8fca4656fe14bc2f63c7966be0877d999" exitCode=0 Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.459906 4775 generic.go:334] "Generic (PLEG): container finished" podID="fb3f1e05-6056-4a7f-9452-22caa26e74fd" containerID="5417129905c56019139e8fe84dd07140ab4c584dde31a376f01ba295d9052432" exitCode=143 Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.459801 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fb3f1e05-6056-4a7f-9452-22caa26e74fd","Type":"ContainerDied","Data":"4fa32c27636fc153967f4aaaf1df73d8fca4656fe14bc2f63c7966be0877d999"} Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.459985 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fb3f1e05-6056-4a7f-9452-22caa26e74fd","Type":"ContainerDied","Data":"5417129905c56019139e8fe84dd07140ab4c584dde31a376f01ba295d9052432"} Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.462189 4775 generic.go:334] "Generic (PLEG): container finished" podID="636228b7-669b-4b5d-abec-bf78cb1513f0" containerID="9a436c521b9cde86255d5d1bb86e7e7617657d67544c27a71666b10506ae93dc" exitCode=143 Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.462245 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"636228b7-669b-4b5d-abec-bf78cb1513f0","Type":"ContainerDied","Data":"9a436c521b9cde86255d5d1bb86e7e7617657d67544c27a71666b10506ae93dc"} Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.464222 4775 generic.go:334] "Generic (PLEG): container finished" podID="33cb4f91-aca0-4221-873f-e2c16e30ccee" containerID="96a2bef185388797d3fea57d8ac4fefdb91c53837dbce4eb9abcae899aab9b04" exitCode=0 Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.464272 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567830-jv78x" event={"ID":"33cb4f91-aca0-4221-873f-e2c16e30ccee","Type":"ContainerDied","Data":"96a2bef185388797d3fea57d8ac4fefdb91c53837dbce4eb9abcae899aab9b04"} Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.861911 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.933395 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-public-tls-certs\") pod \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.933458 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb3f1e05-6056-4a7f-9452-22caa26e74fd-logs\") pod \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.933482 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jdpm\" (UniqueName: \"kubernetes.io/projected/fb3f1e05-6056-4a7f-9452-22caa26e74fd-kube-api-access-8jdpm\") pod \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.933639 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-combined-ca-bundle\") pod \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.933871 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb3f1e05-6056-4a7f-9452-22caa26e74fd-logs" (OuterVolumeSpecName: "logs") pod "fb3f1e05-6056-4a7f-9452-22caa26e74fd" (UID: "fb3f1e05-6056-4a7f-9452-22caa26e74fd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.934175 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-internal-tls-certs\") pod \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.934399 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-config-data\") pod \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\" (UID: \"fb3f1e05-6056-4a7f-9452-22caa26e74fd\") " Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.936025 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb3f1e05-6056-4a7f-9452-22caa26e74fd-logs\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.938730 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb3f1e05-6056-4a7f-9452-22caa26e74fd-kube-api-access-8jdpm" (OuterVolumeSpecName: "kube-api-access-8jdpm") pod "fb3f1e05-6056-4a7f-9452-22caa26e74fd" (UID: "fb3f1e05-6056-4a7f-9452-22caa26e74fd"). InnerVolumeSpecName "kube-api-access-8jdpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.962064 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb3f1e05-6056-4a7f-9452-22caa26e74fd" (UID: "fb3f1e05-6056-4a7f-9452-22caa26e74fd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.971290 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-config-data" (OuterVolumeSpecName: "config-data") pod "fb3f1e05-6056-4a7f-9452-22caa26e74fd" (UID: "fb3f1e05-6056-4a7f-9452-22caa26e74fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.983982 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "fb3f1e05-6056-4a7f-9452-22caa26e74fd" (UID: "fb3f1e05-6056-4a7f-9452-22caa26e74fd"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:10:05 crc kubenswrapper[4775]: I0321 05:10:05.991492 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "fb3f1e05-6056-4a7f-9452-22caa26e74fd" (UID: "fb3f1e05-6056-4a7f-9452-22caa26e74fd"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.037619 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.037924 4775 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.037936 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jdpm\" (UniqueName: \"kubernetes.io/projected/fb3f1e05-6056-4a7f-9452-22caa26e74fd-kube-api-access-8jdpm\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.037945 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.037953 4775 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb3f1e05-6056-4a7f-9452-22caa26e74fd-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:06 crc kubenswrapper[4775]: E0321 05:10:06.374264 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c561875841331ee73e6bcb2e365ce80407bc0539d3ec5ad570942023fedeb9c2 is running failed: container process not found" containerID="c561875841331ee73e6bcb2e365ce80407bc0539d3ec5ad570942023fedeb9c2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 21 05:10:06 crc kubenswrapper[4775]: E0321 05:10:06.374700 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c561875841331ee73e6bcb2e365ce80407bc0539d3ec5ad570942023fedeb9c2 is running failed: container process not found" containerID="c561875841331ee73e6bcb2e365ce80407bc0539d3ec5ad570942023fedeb9c2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 21 05:10:06 crc kubenswrapper[4775]: E0321 05:10:06.375047 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c561875841331ee73e6bcb2e365ce80407bc0539d3ec5ad570942023fedeb9c2 is running failed: container process not found" containerID="c561875841331ee73e6bcb2e365ce80407bc0539d3ec5ad570942023fedeb9c2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 21 05:10:06 crc kubenswrapper[4775]: E0321 05:10:06.375086 4775 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c561875841331ee73e6bcb2e365ce80407bc0539d3ec5ad570942023fedeb9c2 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="0f713501-fa26-4ef2-87ac-c5376b84891d" containerName="nova-scheduler-scheduler" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.477434 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fb3f1e05-6056-4a7f-9452-22caa26e74fd","Type":"ContainerDied","Data":"58500c9c4082d59c6b4aa8467cacee4491bef38fcd23aebf0ec570b68876648c"} Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.477493 4775 scope.go:117] "RemoveContainer" containerID="4fa32c27636fc153967f4aaaf1df73d8fca4656fe14bc2f63c7966be0877d999" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.477643 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.492369 4775 generic.go:334] "Generic (PLEG): container finished" podID="0f713501-fa26-4ef2-87ac-c5376b84891d" containerID="c561875841331ee73e6bcb2e365ce80407bc0539d3ec5ad570942023fedeb9c2" exitCode=0 Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.492479 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0f713501-fa26-4ef2-87ac-c5376b84891d","Type":"ContainerDied","Data":"c561875841331ee73e6bcb2e365ce80407bc0539d3ec5ad570942023fedeb9c2"} Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.498295 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fdf78a6-b107-4949-affa-6152d15afda0","Type":"ContainerStarted","Data":"5c6a6597ad7154fd7a253fa86eb6c5cde8565f20b6949e97002d664b75764604"} Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.525473 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.544103 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.556095 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 21 05:10:06 crc kubenswrapper[4775]: E0321 05:10:06.556619 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb3f1e05-6056-4a7f-9452-22caa26e74fd" containerName="nova-api-log" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.556638 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb3f1e05-6056-4a7f-9452-22caa26e74fd" containerName="nova-api-log" Mar 21 05:10:06 crc kubenswrapper[4775]: E0321 05:10:06.556660 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3172d76-da59-4dee-95df-6b74b7fb0033" containerName="nova-manage" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.556669 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3172d76-da59-4dee-95df-6b74b7fb0033" containerName="nova-manage" Mar 21 05:10:06 crc kubenswrapper[4775]: E0321 05:10:06.556691 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb3f1e05-6056-4a7f-9452-22caa26e74fd" containerName="nova-api-api" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.556700 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb3f1e05-6056-4a7f-9452-22caa26e74fd" containerName="nova-api-api" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.556937 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3172d76-da59-4dee-95df-6b74b7fb0033" containerName="nova-manage" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.556954 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb3f1e05-6056-4a7f-9452-22caa26e74fd" containerName="nova-api-api" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.556975 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb3f1e05-6056-4a7f-9452-22caa26e74fd" containerName="nova-api-log" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.558262 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.560634 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.561722 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.563702 4775 scope.go:117] "RemoveContainer" containerID="5417129905c56019139e8fe84dd07140ab4c584dde31a376f01ba295d9052432" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.573226 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.576492 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.653153 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/886c404c-ceec-48e7-90da-96d6aa201152-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"886c404c-ceec-48e7-90da-96d6aa201152\") " pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.653494 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv8md\" (UniqueName: \"kubernetes.io/projected/886c404c-ceec-48e7-90da-96d6aa201152-kube-api-access-bv8md\") pod \"nova-api-0\" (UID: \"886c404c-ceec-48e7-90da-96d6aa201152\") " pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.653534 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/886c404c-ceec-48e7-90da-96d6aa201152-internal-tls-certs\") pod \"nova-api-0\" (UID: \"886c404c-ceec-48e7-90da-96d6aa201152\") " pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.653569 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/886c404c-ceec-48e7-90da-96d6aa201152-config-data\") pod \"nova-api-0\" (UID: \"886c404c-ceec-48e7-90da-96d6aa201152\") " pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.653585 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/886c404c-ceec-48e7-90da-96d6aa201152-logs\") pod \"nova-api-0\" (UID: \"886c404c-ceec-48e7-90da-96d6aa201152\") " pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.653676 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/886c404c-ceec-48e7-90da-96d6aa201152-public-tls-certs\") pod \"nova-api-0\" (UID: \"886c404c-ceec-48e7-90da-96d6aa201152\") " pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.691262 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.754298 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f713501-fa26-4ef2-87ac-c5376b84891d-config-data\") pod \"0f713501-fa26-4ef2-87ac-c5376b84891d\" (UID: \"0f713501-fa26-4ef2-87ac-c5376b84891d\") " Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.754431 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f713501-fa26-4ef2-87ac-c5376b84891d-combined-ca-bundle\") pod \"0f713501-fa26-4ef2-87ac-c5376b84891d\" (UID: \"0f713501-fa26-4ef2-87ac-c5376b84891d\") " Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.754597 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jkv8\" (UniqueName: \"kubernetes.io/projected/0f713501-fa26-4ef2-87ac-c5376b84891d-kube-api-access-5jkv8\") pod \"0f713501-fa26-4ef2-87ac-c5376b84891d\" (UID: \"0f713501-fa26-4ef2-87ac-c5376b84891d\") " Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.754803 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/886c404c-ceec-48e7-90da-96d6aa201152-internal-tls-certs\") pod \"nova-api-0\" (UID: \"886c404c-ceec-48e7-90da-96d6aa201152\") " pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.754840 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/886c404c-ceec-48e7-90da-96d6aa201152-config-data\") pod \"nova-api-0\" (UID: \"886c404c-ceec-48e7-90da-96d6aa201152\") " pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.754856 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/886c404c-ceec-48e7-90da-96d6aa201152-logs\") pod \"nova-api-0\" (UID: \"886c404c-ceec-48e7-90da-96d6aa201152\") " pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.754919 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/886c404c-ceec-48e7-90da-96d6aa201152-public-tls-certs\") pod \"nova-api-0\" (UID: \"886c404c-ceec-48e7-90da-96d6aa201152\") " pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.754970 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/886c404c-ceec-48e7-90da-96d6aa201152-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"886c404c-ceec-48e7-90da-96d6aa201152\") " pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.755092 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv8md\" (UniqueName: \"kubernetes.io/projected/886c404c-ceec-48e7-90da-96d6aa201152-kube-api-access-bv8md\") pod \"nova-api-0\" (UID: \"886c404c-ceec-48e7-90da-96d6aa201152\") " pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.760614 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/886c404c-ceec-48e7-90da-96d6aa201152-logs\") pod \"nova-api-0\" (UID: \"886c404c-ceec-48e7-90da-96d6aa201152\") " pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.766000 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/886c404c-ceec-48e7-90da-96d6aa201152-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"886c404c-ceec-48e7-90da-96d6aa201152\") " pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.767347 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f713501-fa26-4ef2-87ac-c5376b84891d-kube-api-access-5jkv8" (OuterVolumeSpecName: "kube-api-access-5jkv8") pod "0f713501-fa26-4ef2-87ac-c5376b84891d" (UID: "0f713501-fa26-4ef2-87ac-c5376b84891d"). InnerVolumeSpecName "kube-api-access-5jkv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.769082 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/886c404c-ceec-48e7-90da-96d6aa201152-config-data\") pod \"nova-api-0\" (UID: \"886c404c-ceec-48e7-90da-96d6aa201152\") " pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.770401 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/886c404c-ceec-48e7-90da-96d6aa201152-public-tls-certs\") pod \"nova-api-0\" (UID: \"886c404c-ceec-48e7-90da-96d6aa201152\") " pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.776131 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/886c404c-ceec-48e7-90da-96d6aa201152-internal-tls-certs\") pod \"nova-api-0\" (UID: \"886c404c-ceec-48e7-90da-96d6aa201152\") " pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.796245 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv8md\" (UniqueName: \"kubernetes.io/projected/886c404c-ceec-48e7-90da-96d6aa201152-kube-api-access-bv8md\") pod \"nova-api-0\" (UID: \"886c404c-ceec-48e7-90da-96d6aa201152\") " pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.808711 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f713501-fa26-4ef2-87ac-c5376b84891d-config-data" (OuterVolumeSpecName: "config-data") pod "0f713501-fa26-4ef2-87ac-c5376b84891d" (UID: "0f713501-fa26-4ef2-87ac-c5376b84891d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.838814 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f713501-fa26-4ef2-87ac-c5376b84891d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f713501-fa26-4ef2-87ac-c5376b84891d" (UID: "0f713501-fa26-4ef2-87ac-c5376b84891d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.856327 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jkv8\" (UniqueName: \"kubernetes.io/projected/0f713501-fa26-4ef2-87ac-c5376b84891d-kube-api-access-5jkv8\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.856356 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f713501-fa26-4ef2-87ac-c5376b84891d-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.856365 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f713501-fa26-4ef2-87ac-c5376b84891d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.893931 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.905036 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567830-jv78x" Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.957355 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ns84k\" (UniqueName: \"kubernetes.io/projected/33cb4f91-aca0-4221-873f-e2c16e30ccee-kube-api-access-ns84k\") pod \"33cb4f91-aca0-4221-873f-e2c16e30ccee\" (UID: \"33cb4f91-aca0-4221-873f-e2c16e30ccee\") " Mar 21 05:10:06 crc kubenswrapper[4775]: I0321 05:10:06.961142 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33cb4f91-aca0-4221-873f-e2c16e30ccee-kube-api-access-ns84k" (OuterVolumeSpecName: "kube-api-access-ns84k") pod "33cb4f91-aca0-4221-873f-e2c16e30ccee" (UID: "33cb4f91-aca0-4221-873f-e2c16e30ccee"). InnerVolumeSpecName "kube-api-access-ns84k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.059668 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ns84k\" (UniqueName: \"kubernetes.io/projected/33cb4f91-aca0-4221-873f-e2c16e30ccee-kube-api-access-ns84k\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.491050 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.510813 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0f713501-fa26-4ef2-87ac-c5376b84891d","Type":"ContainerDied","Data":"35d35b17b5e3ddd8509d6d47f432a3246971c8785668db26e3bf835f5d64e720"} Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.510867 4775 scope.go:117] "RemoveContainer" containerID="c561875841331ee73e6bcb2e365ce80407bc0539d3ec5ad570942023fedeb9c2" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.510978 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.516715 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fdf78a6-b107-4949-affa-6152d15afda0","Type":"ContainerStarted","Data":"02056fc0b14ba46d658b8a5aac716fe485d874dd4d3feb41a0e56ea871cdd2a0"} Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.526526 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567830-jv78x" event={"ID":"33cb4f91-aca0-4221-873f-e2c16e30ccee","Type":"ContainerDied","Data":"e13af9720ea7d69a98c04ade641d3957d885af03c65baffd94332be373fb4ff3"} Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.526571 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e13af9720ea7d69a98c04ade641d3957d885af03c65baffd94332be373fb4ff3" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.526625 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567830-jv78x" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.562170 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.585238 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.598664 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 05:10:07 crc kubenswrapper[4775]: E0321 05:10:07.599154 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33cb4f91-aca0-4221-873f-e2c16e30ccee" containerName="oc" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.599170 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="33cb4f91-aca0-4221-873f-e2c16e30ccee" containerName="oc" Mar 21 05:10:07 crc kubenswrapper[4775]: E0321 05:10:07.599211 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f713501-fa26-4ef2-87ac-c5376b84891d" containerName="nova-scheduler-scheduler" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.599229 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f713501-fa26-4ef2-87ac-c5376b84891d" containerName="nova-scheduler-scheduler" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.599527 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="33cb4f91-aca0-4221-873f-e2c16e30ccee" containerName="oc" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.599555 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f713501-fa26-4ef2-87ac-c5376b84891d" containerName="nova-scheduler-scheduler" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.600276 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.603175 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.607399 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.672494 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bab55731-40da-4831-a8b5-f9c413452367-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bab55731-40da-4831-a8b5-f9c413452367\") " pod="openstack/nova-scheduler-0" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.672571 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvqwh\" (UniqueName: \"kubernetes.io/projected/bab55731-40da-4831-a8b5-f9c413452367-kube-api-access-lvqwh\") pod \"nova-scheduler-0\" (UID: \"bab55731-40da-4831-a8b5-f9c413452367\") " pod="openstack/nova-scheduler-0" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.672662 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bab55731-40da-4831-a8b5-f9c413452367-config-data\") pod \"nova-scheduler-0\" (UID: \"bab55731-40da-4831-a8b5-f9c413452367\") " pod="openstack/nova-scheduler-0" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.688902 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f713501-fa26-4ef2-87ac-c5376b84891d" path="/var/lib/kubelet/pods/0f713501-fa26-4ef2-87ac-c5376b84891d/volumes" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.689681 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb3f1e05-6056-4a7f-9452-22caa26e74fd" path="/var/lib/kubelet/pods/fb3f1e05-6056-4a7f-9452-22caa26e74fd/volumes" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.774869 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bab55731-40da-4831-a8b5-f9c413452367-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bab55731-40da-4831-a8b5-f9c413452367\") " pod="openstack/nova-scheduler-0" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.775214 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvqwh\" (UniqueName: \"kubernetes.io/projected/bab55731-40da-4831-a8b5-f9c413452367-kube-api-access-lvqwh\") pod \"nova-scheduler-0\" (UID: \"bab55731-40da-4831-a8b5-f9c413452367\") " pod="openstack/nova-scheduler-0" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.775244 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bab55731-40da-4831-a8b5-f9c413452367-config-data\") pod \"nova-scheduler-0\" (UID: \"bab55731-40da-4831-a8b5-f9c413452367\") " pod="openstack/nova-scheduler-0" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.780776 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bab55731-40da-4831-a8b5-f9c413452367-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bab55731-40da-4831-a8b5-f9c413452367\") " pod="openstack/nova-scheduler-0" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.783021 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bab55731-40da-4831-a8b5-f9c413452367-config-data\") pod \"nova-scheduler-0\" (UID: \"bab55731-40da-4831-a8b5-f9c413452367\") " pod="openstack/nova-scheduler-0" Mar 21 05:10:07 crc kubenswrapper[4775]: I0321 05:10:07.799633 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvqwh\" (UniqueName: \"kubernetes.io/projected/bab55731-40da-4831-a8b5-f9c413452367-kube-api-access-lvqwh\") pod \"nova-scheduler-0\" (UID: \"bab55731-40da-4831-a8b5-f9c413452367\") " pod="openstack/nova-scheduler-0" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.007005 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567824-jsc7j"] Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.014252 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567824-jsc7j"] Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.051177 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.239471 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.284311 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636228b7-669b-4b5d-abec-bf78cb1513f0-combined-ca-bundle\") pod \"636228b7-669b-4b5d-abec-bf78cb1513f0\" (UID: \"636228b7-669b-4b5d-abec-bf78cb1513f0\") " Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.284425 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/636228b7-669b-4b5d-abec-bf78cb1513f0-logs\") pod \"636228b7-669b-4b5d-abec-bf78cb1513f0\" (UID: \"636228b7-669b-4b5d-abec-bf78cb1513f0\") " Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.284581 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636228b7-669b-4b5d-abec-bf78cb1513f0-config-data\") pod \"636228b7-669b-4b5d-abec-bf78cb1513f0\" (UID: \"636228b7-669b-4b5d-abec-bf78cb1513f0\") " Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.284649 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/636228b7-669b-4b5d-abec-bf78cb1513f0-nova-metadata-tls-certs\") pod \"636228b7-669b-4b5d-abec-bf78cb1513f0\" (UID: \"636228b7-669b-4b5d-abec-bf78cb1513f0\") " Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.284738 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plq2h\" (UniqueName: \"kubernetes.io/projected/636228b7-669b-4b5d-abec-bf78cb1513f0-kube-api-access-plq2h\") pod \"636228b7-669b-4b5d-abec-bf78cb1513f0\" (UID: \"636228b7-669b-4b5d-abec-bf78cb1513f0\") " Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.286441 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/636228b7-669b-4b5d-abec-bf78cb1513f0-logs" (OuterVolumeSpecName: "logs") pod "636228b7-669b-4b5d-abec-bf78cb1513f0" (UID: "636228b7-669b-4b5d-abec-bf78cb1513f0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.292385 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/636228b7-669b-4b5d-abec-bf78cb1513f0-kube-api-access-plq2h" (OuterVolumeSpecName: "kube-api-access-plq2h") pod "636228b7-669b-4b5d-abec-bf78cb1513f0" (UID: "636228b7-669b-4b5d-abec-bf78cb1513f0"). InnerVolumeSpecName "kube-api-access-plq2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.336261 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/636228b7-669b-4b5d-abec-bf78cb1513f0-config-data" (OuterVolumeSpecName: "config-data") pod "636228b7-669b-4b5d-abec-bf78cb1513f0" (UID: "636228b7-669b-4b5d-abec-bf78cb1513f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.368426 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/636228b7-669b-4b5d-abec-bf78cb1513f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "636228b7-669b-4b5d-abec-bf78cb1513f0" (UID: "636228b7-669b-4b5d-abec-bf78cb1513f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.387997 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/636228b7-669b-4b5d-abec-bf78cb1513f0-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "636228b7-669b-4b5d-abec-bf78cb1513f0" (UID: "636228b7-669b-4b5d-abec-bf78cb1513f0"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.390491 4775 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/636228b7-669b-4b5d-abec-bf78cb1513f0-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.390515 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plq2h\" (UniqueName: \"kubernetes.io/projected/636228b7-669b-4b5d-abec-bf78cb1513f0-kube-api-access-plq2h\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.390524 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636228b7-669b-4b5d-abec-bf78cb1513f0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.390533 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/636228b7-669b-4b5d-abec-bf78cb1513f0-logs\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.390542 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636228b7-669b-4b5d-abec-bf78cb1513f0-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.531746 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.543049 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"886c404c-ceec-48e7-90da-96d6aa201152","Type":"ContainerStarted","Data":"e15bb364282c0ba2931783f4dafe88fe9283770f5c4f7fa427724353592f1c45"} Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.543101 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"886c404c-ceec-48e7-90da-96d6aa201152","Type":"ContainerStarted","Data":"01cbfe40ccd945f67c2bba82115e395d0b9610b14660c86f3c227e8c213758b1"} Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.543130 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"886c404c-ceec-48e7-90da-96d6aa201152","Type":"ContainerStarted","Data":"2d7a24379beecfe6c23a41760c2ac92eb37fa2a9f779e013f2c6319dd30ffc87"} Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.550056 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bab55731-40da-4831-a8b5-f9c413452367","Type":"ContainerStarted","Data":"54239add11ebd7fbe13885f3881605227b4aa3474202a89dd331c69292908f27"} Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.552184 4775 generic.go:334] "Generic (PLEG): container finished" podID="636228b7-669b-4b5d-abec-bf78cb1513f0" containerID="88b942c0d478eda3ab139d99845879b76d209947d510208b64a88f8f909afefc" exitCode=0 Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.552239 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"636228b7-669b-4b5d-abec-bf78cb1513f0","Type":"ContainerDied","Data":"88b942c0d478eda3ab139d99845879b76d209947d510208b64a88f8f909afefc"} Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.552257 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"636228b7-669b-4b5d-abec-bf78cb1513f0","Type":"ContainerDied","Data":"c6bcff38ed2ee79c0028a43362b2200b49c362342fd98956493476b35ddff899"} Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.552273 4775 scope.go:117] "RemoveContainer" containerID="88b942c0d478eda3ab139d99845879b76d209947d510208b64a88f8f909afefc" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.552547 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.588653 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.588614888 podStartE2EDuration="2.588614888s" podCreationTimestamp="2026-03-21 05:10:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:10:08.568620503 +0000 UTC m=+1361.545084127" watchObservedRunningTime="2026-03-21 05:10:08.588614888 +0000 UTC m=+1361.565078512" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.707514 4775 scope.go:117] "RemoveContainer" containerID="9a436c521b9cde86255d5d1bb86e7e7617657d67544c27a71666b10506ae93dc" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.748826 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.749896 4775 scope.go:117] "RemoveContainer" containerID="88b942c0d478eda3ab139d99845879b76d209947d510208b64a88f8f909afefc" Mar 21 05:10:08 crc kubenswrapper[4775]: E0321 05:10:08.752524 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88b942c0d478eda3ab139d99845879b76d209947d510208b64a88f8f909afefc\": container with ID starting with 88b942c0d478eda3ab139d99845879b76d209947d510208b64a88f8f909afefc not found: ID does not exist" containerID="88b942c0d478eda3ab139d99845879b76d209947d510208b64a88f8f909afefc" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.752568 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88b942c0d478eda3ab139d99845879b76d209947d510208b64a88f8f909afefc"} err="failed to get container status \"88b942c0d478eda3ab139d99845879b76d209947d510208b64a88f8f909afefc\": rpc error: code = NotFound desc = could not find container \"88b942c0d478eda3ab139d99845879b76d209947d510208b64a88f8f909afefc\": container with ID starting with 88b942c0d478eda3ab139d99845879b76d209947d510208b64a88f8f909afefc not found: ID does not exist" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.752595 4775 scope.go:117] "RemoveContainer" containerID="9a436c521b9cde86255d5d1bb86e7e7617657d67544c27a71666b10506ae93dc" Mar 21 05:10:08 crc kubenswrapper[4775]: E0321 05:10:08.757226 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a436c521b9cde86255d5d1bb86e7e7617657d67544c27a71666b10506ae93dc\": container with ID starting with 9a436c521b9cde86255d5d1bb86e7e7617657d67544c27a71666b10506ae93dc not found: ID does not exist" containerID="9a436c521b9cde86255d5d1bb86e7e7617657d67544c27a71666b10506ae93dc" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.757267 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a436c521b9cde86255d5d1bb86e7e7617657d67544c27a71666b10506ae93dc"} err="failed to get container status \"9a436c521b9cde86255d5d1bb86e7e7617657d67544c27a71666b10506ae93dc\": rpc error: code = NotFound desc = could not find container \"9a436c521b9cde86255d5d1bb86e7e7617657d67544c27a71666b10506ae93dc\": container with ID starting with 9a436c521b9cde86255d5d1bb86e7e7617657d67544c27a71666b10506ae93dc not found: ID does not exist" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.762184 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.770051 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:10:08 crc kubenswrapper[4775]: E0321 05:10:08.770451 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="636228b7-669b-4b5d-abec-bf78cb1513f0" containerName="nova-metadata-metadata" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.770467 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="636228b7-669b-4b5d-abec-bf78cb1513f0" containerName="nova-metadata-metadata" Mar 21 05:10:08 crc kubenswrapper[4775]: E0321 05:10:08.770481 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="636228b7-669b-4b5d-abec-bf78cb1513f0" containerName="nova-metadata-log" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.770487 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="636228b7-669b-4b5d-abec-bf78cb1513f0" containerName="nova-metadata-log" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.770655 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="636228b7-669b-4b5d-abec-bf78cb1513f0" containerName="nova-metadata-metadata" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.770676 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="636228b7-669b-4b5d-abec-bf78cb1513f0" containerName="nova-metadata-log" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.771904 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.775156 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.790307 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.800246 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnlxt\" (UniqueName: \"kubernetes.io/projected/208cfa71-8242-4958-b9db-21fc180a6697-kube-api-access-tnlxt\") pod \"nova-metadata-0\" (UID: \"208cfa71-8242-4958-b9db-21fc180a6697\") " pod="openstack/nova-metadata-0" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.800342 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/208cfa71-8242-4958-b9db-21fc180a6697-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"208cfa71-8242-4958-b9db-21fc180a6697\") " pod="openstack/nova-metadata-0" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.800388 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/208cfa71-8242-4958-b9db-21fc180a6697-logs\") pod \"nova-metadata-0\" (UID: \"208cfa71-8242-4958-b9db-21fc180a6697\") " pod="openstack/nova-metadata-0" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.800430 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/208cfa71-8242-4958-b9db-21fc180a6697-config-data\") pod \"nova-metadata-0\" (UID: \"208cfa71-8242-4958-b9db-21fc180a6697\") " pod="openstack/nova-metadata-0" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.800456 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/208cfa71-8242-4958-b9db-21fc180a6697-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"208cfa71-8242-4958-b9db-21fc180a6697\") " pod="openstack/nova-metadata-0" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.804863 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.902232 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnlxt\" (UniqueName: \"kubernetes.io/projected/208cfa71-8242-4958-b9db-21fc180a6697-kube-api-access-tnlxt\") pod \"nova-metadata-0\" (UID: \"208cfa71-8242-4958-b9db-21fc180a6697\") " pod="openstack/nova-metadata-0" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.902337 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/208cfa71-8242-4958-b9db-21fc180a6697-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"208cfa71-8242-4958-b9db-21fc180a6697\") " pod="openstack/nova-metadata-0" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.902390 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/208cfa71-8242-4958-b9db-21fc180a6697-logs\") pod \"nova-metadata-0\" (UID: \"208cfa71-8242-4958-b9db-21fc180a6697\") " pod="openstack/nova-metadata-0" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.902433 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/208cfa71-8242-4958-b9db-21fc180a6697-config-data\") pod \"nova-metadata-0\" (UID: \"208cfa71-8242-4958-b9db-21fc180a6697\") " pod="openstack/nova-metadata-0" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.902463 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/208cfa71-8242-4958-b9db-21fc180a6697-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"208cfa71-8242-4958-b9db-21fc180a6697\") " pod="openstack/nova-metadata-0" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.902904 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/208cfa71-8242-4958-b9db-21fc180a6697-logs\") pod \"nova-metadata-0\" (UID: \"208cfa71-8242-4958-b9db-21fc180a6697\") " pod="openstack/nova-metadata-0" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.907491 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/208cfa71-8242-4958-b9db-21fc180a6697-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"208cfa71-8242-4958-b9db-21fc180a6697\") " pod="openstack/nova-metadata-0" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.907525 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/208cfa71-8242-4958-b9db-21fc180a6697-config-data\") pod \"nova-metadata-0\" (UID: \"208cfa71-8242-4958-b9db-21fc180a6697\") " pod="openstack/nova-metadata-0" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.909601 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/208cfa71-8242-4958-b9db-21fc180a6697-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"208cfa71-8242-4958-b9db-21fc180a6697\") " pod="openstack/nova-metadata-0" Mar 21 05:10:08 crc kubenswrapper[4775]: I0321 05:10:08.923506 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnlxt\" (UniqueName: \"kubernetes.io/projected/208cfa71-8242-4958-b9db-21fc180a6697-kube-api-access-tnlxt\") pod \"nova-metadata-0\" (UID: \"208cfa71-8242-4958-b9db-21fc180a6697\") " pod="openstack/nova-metadata-0" Mar 21 05:10:09 crc kubenswrapper[4775]: I0321 05:10:09.116649 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 05:10:09 crc kubenswrapper[4775]: I0321 05:10:09.560858 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 05:10:09 crc kubenswrapper[4775]: I0321 05:10:09.578311 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"208cfa71-8242-4958-b9db-21fc180a6697","Type":"ContainerStarted","Data":"bef23ce3537e304a8440703679a26f74fdb34f50019dcbd73a0a31c53033d754"} Mar 21 05:10:09 crc kubenswrapper[4775]: I0321 05:10:09.581382 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bab55731-40da-4831-a8b5-f9c413452367","Type":"ContainerStarted","Data":"1bf60e4668a1f23910cdef9ea4cd616c9ffe8afc8a12261f33ae430c857beb32"} Mar 21 05:10:09 crc kubenswrapper[4775]: I0321 05:10:09.590178 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fdf78a6-b107-4949-affa-6152d15afda0","Type":"ContainerStarted","Data":"ef9369d555365416b278955eec54a4b9a28ad97b7c0e8c0f39d301bd6d2b5a92"} Mar 21 05:10:09 crc kubenswrapper[4775]: I0321 05:10:09.590683 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 21 05:10:09 crc kubenswrapper[4775]: I0321 05:10:09.603279 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.603256831 podStartE2EDuration="2.603256831s" podCreationTimestamp="2026-03-21 05:10:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:10:09.595892253 +0000 UTC m=+1362.572355887" watchObservedRunningTime="2026-03-21 05:10:09.603256831 +0000 UTC m=+1362.579720465" Mar 21 05:10:09 crc kubenswrapper[4775]: I0321 05:10:09.647154 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.956140758 podStartE2EDuration="9.647132452s" podCreationTimestamp="2026-03-21 05:10:00 +0000 UTC" firstStartedPulling="2026-03-21 05:10:01.330208553 +0000 UTC m=+1354.306672177" lastFinishedPulling="2026-03-21 05:10:09.021200247 +0000 UTC m=+1361.997663871" observedRunningTime="2026-03-21 05:10:09.639390993 +0000 UTC m=+1362.615854617" watchObservedRunningTime="2026-03-21 05:10:09.647132452 +0000 UTC m=+1362.623596076" Mar 21 05:10:09 crc kubenswrapper[4775]: I0321 05:10:09.677708 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="636228b7-669b-4b5d-abec-bf78cb1513f0" path="/var/lib/kubelet/pods/636228b7-669b-4b5d-abec-bf78cb1513f0/volumes" Mar 21 05:10:09 crc kubenswrapper[4775]: I0321 05:10:09.678261 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c861c77c-15ba-4204-8603-e6093bc1a0b8" path="/var/lib/kubelet/pods/c861c77c-15ba-4204-8603-e6093bc1a0b8/volumes" Mar 21 05:10:10 crc kubenswrapper[4775]: I0321 05:10:10.602850 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"208cfa71-8242-4958-b9db-21fc180a6697","Type":"ContainerStarted","Data":"710f8b630e8fd4d943202a35b0f5180cd356bcfa3ee0c35b8ed1b0b486c93e15"} Mar 21 05:10:11 crc kubenswrapper[4775]: I0321 05:10:11.613559 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"208cfa71-8242-4958-b9db-21fc180a6697","Type":"ContainerStarted","Data":"f6c428a9e20773a6c9af646402b83b649bee9c253793ddd00227e15c99dfbb48"} Mar 21 05:10:11 crc kubenswrapper[4775]: I0321 05:10:11.636216 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.636178672 podStartE2EDuration="3.636178672s" podCreationTimestamp="2026-03-21 05:10:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:10:11.629594246 +0000 UTC m=+1364.606057890" watchObservedRunningTime="2026-03-21 05:10:11.636178672 +0000 UTC m=+1364.612642316" Mar 21 05:10:13 crc kubenswrapper[4775]: I0321 05:10:13.052030 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 21 05:10:16 crc kubenswrapper[4775]: I0321 05:10:16.894263 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 21 05:10:16 crc kubenswrapper[4775]: I0321 05:10:16.894754 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 21 05:10:17 crc kubenswrapper[4775]: I0321 05:10:17.909364 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="886c404c-ceec-48e7-90da-96d6aa201152" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.206:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 21 05:10:17 crc kubenswrapper[4775]: I0321 05:10:17.910936 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="886c404c-ceec-48e7-90da-96d6aa201152" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.206:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 21 05:10:18 crc kubenswrapper[4775]: I0321 05:10:18.052778 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 21 05:10:18 crc kubenswrapper[4775]: I0321 05:10:18.078222 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 21 05:10:18 crc kubenswrapper[4775]: I0321 05:10:18.727277 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 21 05:10:19 crc kubenswrapper[4775]: I0321 05:10:19.117720 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 21 05:10:19 crc kubenswrapper[4775]: I0321 05:10:19.117837 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 21 05:10:20 crc kubenswrapper[4775]: I0321 05:10:20.128514 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="208cfa71-8242-4958-b9db-21fc180a6697" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.208:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 21 05:10:20 crc kubenswrapper[4775]: I0321 05:10:20.128551 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="208cfa71-8242-4958-b9db-21fc180a6697" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.208:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 21 05:10:20 crc kubenswrapper[4775]: I0321 05:10:20.568521 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="e3dd1c1b-7aba-4083-8a9c-037cd38b2d53" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.194:3000/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 21 05:10:24 crc kubenswrapper[4775]: I0321 05:10:24.894566 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 21 05:10:24 crc kubenswrapper[4775]: I0321 05:10:24.895155 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 21 05:10:27 crc kubenswrapper[4775]: I0321 05:10:27.000513 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 21 05:10:27 crc kubenswrapper[4775]: I0321 05:10:27.003113 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 21 05:10:27 crc kubenswrapper[4775]: I0321 05:10:27.005886 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 21 05:10:27 crc kubenswrapper[4775]: I0321 05:10:27.117620 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 21 05:10:27 crc kubenswrapper[4775]: I0321 05:10:27.117769 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 21 05:10:27 crc kubenswrapper[4775]: I0321 05:10:27.804705 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 21 05:10:29 crc kubenswrapper[4775]: I0321 05:10:29.124876 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 21 05:10:29 crc kubenswrapper[4775]: I0321 05:10:29.126454 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 21 05:10:29 crc kubenswrapper[4775]: I0321 05:10:29.130227 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 21 05:10:29 crc kubenswrapper[4775]: I0321 05:10:29.130880 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 21 05:10:31 crc kubenswrapper[4775]: I0321 05:10:31.192998 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 21 05:10:32 crc kubenswrapper[4775]: I0321 05:10:32.482907 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:10:32 crc kubenswrapper[4775]: I0321 05:10:32.483324 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:10:34 crc kubenswrapper[4775]: I0321 05:10:34.870562 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 21 05:10:34 crc kubenswrapper[4775]: I0321 05:10:34.871160 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="31578b15-f84b-4862-ae52-6720dac8f5e2" containerName="kube-state-metrics" containerID="cri-o://ebdac31be6ec007bb4ed96d43a4d78aab3ca618e5b3cc6ddb93c2810c9ec401a" gracePeriod=30 Mar 21 05:10:35 crc kubenswrapper[4775]: I0321 05:10:35.398246 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 21 05:10:35 crc kubenswrapper[4775]: I0321 05:10:35.496552 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6zwf\" (UniqueName: \"kubernetes.io/projected/31578b15-f84b-4862-ae52-6720dac8f5e2-kube-api-access-l6zwf\") pod \"31578b15-f84b-4862-ae52-6720dac8f5e2\" (UID: \"31578b15-f84b-4862-ae52-6720dac8f5e2\") " Mar 21 05:10:35 crc kubenswrapper[4775]: I0321 05:10:35.515349 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31578b15-f84b-4862-ae52-6720dac8f5e2-kube-api-access-l6zwf" (OuterVolumeSpecName: "kube-api-access-l6zwf") pod "31578b15-f84b-4862-ae52-6720dac8f5e2" (UID: "31578b15-f84b-4862-ae52-6720dac8f5e2"). InnerVolumeSpecName "kube-api-access-l6zwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:10:35 crc kubenswrapper[4775]: I0321 05:10:35.599491 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6zwf\" (UniqueName: \"kubernetes.io/projected/31578b15-f84b-4862-ae52-6720dac8f5e2-kube-api-access-l6zwf\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:35 crc kubenswrapper[4775]: I0321 05:10:35.895732 4775 generic.go:334] "Generic (PLEG): container finished" podID="31578b15-f84b-4862-ae52-6720dac8f5e2" containerID="ebdac31be6ec007bb4ed96d43a4d78aab3ca618e5b3cc6ddb93c2810c9ec401a" exitCode=2 Mar 21 05:10:35 crc kubenswrapper[4775]: I0321 05:10:35.895768 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"31578b15-f84b-4862-ae52-6720dac8f5e2","Type":"ContainerDied","Data":"ebdac31be6ec007bb4ed96d43a4d78aab3ca618e5b3cc6ddb93c2810c9ec401a"} Mar 21 05:10:35 crc kubenswrapper[4775]: I0321 05:10:35.895794 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"31578b15-f84b-4862-ae52-6720dac8f5e2","Type":"ContainerDied","Data":"95eba21f713625c8d0e5f46fb846b1edb0522a4501294e1679ea681e44f67781"} Mar 21 05:10:35 crc kubenswrapper[4775]: I0321 05:10:35.895811 4775 scope.go:117] "RemoveContainer" containerID="ebdac31be6ec007bb4ed96d43a4d78aab3ca618e5b3cc6ddb93c2810c9ec401a" Mar 21 05:10:35 crc kubenswrapper[4775]: I0321 05:10:35.895815 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 21 05:10:35 crc kubenswrapper[4775]: I0321 05:10:35.920300 4775 scope.go:117] "RemoveContainer" containerID="ebdac31be6ec007bb4ed96d43a4d78aab3ca618e5b3cc6ddb93c2810c9ec401a" Mar 21 05:10:35 crc kubenswrapper[4775]: E0321 05:10:35.920681 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebdac31be6ec007bb4ed96d43a4d78aab3ca618e5b3cc6ddb93c2810c9ec401a\": container with ID starting with ebdac31be6ec007bb4ed96d43a4d78aab3ca618e5b3cc6ddb93c2810c9ec401a not found: ID does not exist" containerID="ebdac31be6ec007bb4ed96d43a4d78aab3ca618e5b3cc6ddb93c2810c9ec401a" Mar 21 05:10:35 crc kubenswrapper[4775]: I0321 05:10:35.920718 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebdac31be6ec007bb4ed96d43a4d78aab3ca618e5b3cc6ddb93c2810c9ec401a"} err="failed to get container status \"ebdac31be6ec007bb4ed96d43a4d78aab3ca618e5b3cc6ddb93c2810c9ec401a\": rpc error: code = NotFound desc = could not find container \"ebdac31be6ec007bb4ed96d43a4d78aab3ca618e5b3cc6ddb93c2810c9ec401a\": container with ID starting with ebdac31be6ec007bb4ed96d43a4d78aab3ca618e5b3cc6ddb93c2810c9ec401a not found: ID does not exist" Mar 21 05:10:35 crc kubenswrapper[4775]: I0321 05:10:35.923137 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 21 05:10:35 crc kubenswrapper[4775]: I0321 05:10:35.935533 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 21 05:10:35 crc kubenswrapper[4775]: I0321 05:10:35.949000 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 21 05:10:35 crc kubenswrapper[4775]: E0321 05:10:35.949538 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31578b15-f84b-4862-ae52-6720dac8f5e2" containerName="kube-state-metrics" Mar 21 05:10:35 crc kubenswrapper[4775]: I0321 05:10:35.949560 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="31578b15-f84b-4862-ae52-6720dac8f5e2" containerName="kube-state-metrics" Mar 21 05:10:35 crc kubenswrapper[4775]: I0321 05:10:35.949801 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="31578b15-f84b-4862-ae52-6720dac8f5e2" containerName="kube-state-metrics" Mar 21 05:10:35 crc kubenswrapper[4775]: I0321 05:10:35.950581 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 21 05:10:35 crc kubenswrapper[4775]: I0321 05:10:35.952939 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Mar 21 05:10:35 crc kubenswrapper[4775]: I0321 05:10:35.953289 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Mar 21 05:10:35 crc kubenswrapper[4775]: I0321 05:10:35.982218 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 21 05:10:36 crc kubenswrapper[4775]: I0321 05:10:36.107864 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee3b7a0-9eb3-4702-8fb7-3286df60b21b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"7ee3b7a0-9eb3-4702-8fb7-3286df60b21b\") " pod="openstack/kube-state-metrics-0" Mar 21 05:10:36 crc kubenswrapper[4775]: I0321 05:10:36.107934 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/7ee3b7a0-9eb3-4702-8fb7-3286df60b21b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"7ee3b7a0-9eb3-4702-8fb7-3286df60b21b\") " pod="openstack/kube-state-metrics-0" Mar 21 05:10:36 crc kubenswrapper[4775]: I0321 05:10:36.107966 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ee3b7a0-9eb3-4702-8fb7-3286df60b21b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"7ee3b7a0-9eb3-4702-8fb7-3286df60b21b\") " pod="openstack/kube-state-metrics-0" Mar 21 05:10:36 crc kubenswrapper[4775]: I0321 05:10:36.107996 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzt24\" (UniqueName: \"kubernetes.io/projected/7ee3b7a0-9eb3-4702-8fb7-3286df60b21b-kube-api-access-nzt24\") pod \"kube-state-metrics-0\" (UID: \"7ee3b7a0-9eb3-4702-8fb7-3286df60b21b\") " pod="openstack/kube-state-metrics-0" Mar 21 05:10:36 crc kubenswrapper[4775]: I0321 05:10:36.210916 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee3b7a0-9eb3-4702-8fb7-3286df60b21b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"7ee3b7a0-9eb3-4702-8fb7-3286df60b21b\") " pod="openstack/kube-state-metrics-0" Mar 21 05:10:36 crc kubenswrapper[4775]: I0321 05:10:36.211350 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/7ee3b7a0-9eb3-4702-8fb7-3286df60b21b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"7ee3b7a0-9eb3-4702-8fb7-3286df60b21b\") " pod="openstack/kube-state-metrics-0" Mar 21 05:10:36 crc kubenswrapper[4775]: I0321 05:10:36.211419 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ee3b7a0-9eb3-4702-8fb7-3286df60b21b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"7ee3b7a0-9eb3-4702-8fb7-3286df60b21b\") " pod="openstack/kube-state-metrics-0" Mar 21 05:10:36 crc kubenswrapper[4775]: I0321 05:10:36.211479 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzt24\" (UniqueName: \"kubernetes.io/projected/7ee3b7a0-9eb3-4702-8fb7-3286df60b21b-kube-api-access-nzt24\") pod \"kube-state-metrics-0\" (UID: \"7ee3b7a0-9eb3-4702-8fb7-3286df60b21b\") " pod="openstack/kube-state-metrics-0" Mar 21 05:10:36 crc kubenswrapper[4775]: I0321 05:10:36.217304 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/7ee3b7a0-9eb3-4702-8fb7-3286df60b21b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"7ee3b7a0-9eb3-4702-8fb7-3286df60b21b\") " pod="openstack/kube-state-metrics-0" Mar 21 05:10:36 crc kubenswrapper[4775]: I0321 05:10:36.220358 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee3b7a0-9eb3-4702-8fb7-3286df60b21b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"7ee3b7a0-9eb3-4702-8fb7-3286df60b21b\") " pod="openstack/kube-state-metrics-0" Mar 21 05:10:36 crc kubenswrapper[4775]: I0321 05:10:36.221208 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ee3b7a0-9eb3-4702-8fb7-3286df60b21b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"7ee3b7a0-9eb3-4702-8fb7-3286df60b21b\") " pod="openstack/kube-state-metrics-0" Mar 21 05:10:36 crc kubenswrapper[4775]: I0321 05:10:36.236783 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzt24\" (UniqueName: \"kubernetes.io/projected/7ee3b7a0-9eb3-4702-8fb7-3286df60b21b-kube-api-access-nzt24\") pod \"kube-state-metrics-0\" (UID: \"7ee3b7a0-9eb3-4702-8fb7-3286df60b21b\") " pod="openstack/kube-state-metrics-0" Mar 21 05:10:36 crc kubenswrapper[4775]: I0321 05:10:36.265212 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 21 05:10:37 crc kubenswrapper[4775]: I0321 05:10:36.731450 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:10:37 crc kubenswrapper[4775]: I0321 05:10:36.732387 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7fdf78a6-b107-4949-affa-6152d15afda0" containerName="ceilometer-central-agent" containerID="cri-o://8f8cabcb03903e4d89761a18703dd5c97870cd2e3d0ef375a4910ff30a8d99cf" gracePeriod=30 Mar 21 05:10:37 crc kubenswrapper[4775]: I0321 05:10:36.732834 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7fdf78a6-b107-4949-affa-6152d15afda0" containerName="proxy-httpd" containerID="cri-o://ef9369d555365416b278955eec54a4b9a28ad97b7c0e8c0f39d301bd6d2b5a92" gracePeriod=30 Mar 21 05:10:37 crc kubenswrapper[4775]: I0321 05:10:36.732883 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7fdf78a6-b107-4949-affa-6152d15afda0" containerName="sg-core" containerID="cri-o://02056fc0b14ba46d658b8a5aac716fe485d874dd4d3feb41a0e56ea871cdd2a0" gracePeriod=30 Mar 21 05:10:37 crc kubenswrapper[4775]: I0321 05:10:36.732916 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7fdf78a6-b107-4949-affa-6152d15afda0" containerName="ceilometer-notification-agent" containerID="cri-o://5c6a6597ad7154fd7a253fa86eb6c5cde8565f20b6949e97002d664b75764604" gracePeriod=30 Mar 21 05:10:37 crc kubenswrapper[4775]: I0321 05:10:36.759628 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 21 05:10:37 crc kubenswrapper[4775]: I0321 05:10:36.919231 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7ee3b7a0-9eb3-4702-8fb7-3286df60b21b","Type":"ContainerStarted","Data":"e6fcb40947e0a44da02247b3001dd7bbad982138ad811a2a9dff2ffa3dc5b034"} Mar 21 05:10:37 crc kubenswrapper[4775]: I0321 05:10:36.922044 4775 generic.go:334] "Generic (PLEG): container finished" podID="7fdf78a6-b107-4949-affa-6152d15afda0" containerID="02056fc0b14ba46d658b8a5aac716fe485d874dd4d3feb41a0e56ea871cdd2a0" exitCode=2 Mar 21 05:10:37 crc kubenswrapper[4775]: I0321 05:10:36.922146 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fdf78a6-b107-4949-affa-6152d15afda0","Type":"ContainerDied","Data":"02056fc0b14ba46d658b8a5aac716fe485d874dd4d3feb41a0e56ea871cdd2a0"} Mar 21 05:10:37 crc kubenswrapper[4775]: I0321 05:10:37.679311 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31578b15-f84b-4862-ae52-6720dac8f5e2" path="/var/lib/kubelet/pods/31578b15-f84b-4862-ae52-6720dac8f5e2/volumes" Mar 21 05:10:37 crc kubenswrapper[4775]: I0321 05:10:37.937657 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7ee3b7a0-9eb3-4702-8fb7-3286df60b21b","Type":"ContainerStarted","Data":"3e25aa26aea86ebe53beaeba095f2cca5eae987565371ea0d1d0d037281bc4a1"} Mar 21 05:10:37 crc kubenswrapper[4775]: I0321 05:10:37.938095 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 21 05:10:37 crc kubenswrapper[4775]: I0321 05:10:37.941978 4775 generic.go:334] "Generic (PLEG): container finished" podID="7fdf78a6-b107-4949-affa-6152d15afda0" containerID="ef9369d555365416b278955eec54a4b9a28ad97b7c0e8c0f39d301bd6d2b5a92" exitCode=0 Mar 21 05:10:37 crc kubenswrapper[4775]: I0321 05:10:37.942010 4775 generic.go:334] "Generic (PLEG): container finished" podID="7fdf78a6-b107-4949-affa-6152d15afda0" containerID="5c6a6597ad7154fd7a253fa86eb6c5cde8565f20b6949e97002d664b75764604" exitCode=0 Mar 21 05:10:37 crc kubenswrapper[4775]: I0321 05:10:37.942018 4775 generic.go:334] "Generic (PLEG): container finished" podID="7fdf78a6-b107-4949-affa-6152d15afda0" containerID="8f8cabcb03903e4d89761a18703dd5c97870cd2e3d0ef375a4910ff30a8d99cf" exitCode=0 Mar 21 05:10:37 crc kubenswrapper[4775]: I0321 05:10:37.942040 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fdf78a6-b107-4949-affa-6152d15afda0","Type":"ContainerDied","Data":"ef9369d555365416b278955eec54a4b9a28ad97b7c0e8c0f39d301bd6d2b5a92"} Mar 21 05:10:37 crc kubenswrapper[4775]: I0321 05:10:37.942076 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fdf78a6-b107-4949-affa-6152d15afda0","Type":"ContainerDied","Data":"5c6a6597ad7154fd7a253fa86eb6c5cde8565f20b6949e97002d664b75764604"} Mar 21 05:10:37 crc kubenswrapper[4775]: I0321 05:10:37.942098 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fdf78a6-b107-4949-affa-6152d15afda0","Type":"ContainerDied","Data":"8f8cabcb03903e4d89761a18703dd5c97870cd2e3d0ef375a4910ff30a8d99cf"} Mar 21 05:10:37 crc kubenswrapper[4775]: I0321 05:10:37.964045 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.35142522 podStartE2EDuration="2.964019698s" podCreationTimestamp="2026-03-21 05:10:35 +0000 UTC" firstStartedPulling="2026-03-21 05:10:36.766468843 +0000 UTC m=+1389.742932467" lastFinishedPulling="2026-03-21 05:10:37.379063321 +0000 UTC m=+1390.355526945" observedRunningTime="2026-03-21 05:10:37.955390244 +0000 UTC m=+1390.931853878" watchObservedRunningTime="2026-03-21 05:10:37.964019698 +0000 UTC m=+1390.940483332" Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.088446 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.183482 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fdf78a6-b107-4949-affa-6152d15afda0-run-httpd\") pod \"7fdf78a6-b107-4949-affa-6152d15afda0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.183589 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-combined-ca-bundle\") pod \"7fdf78a6-b107-4949-affa-6152d15afda0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.183662 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-config-data\") pod \"7fdf78a6-b107-4949-affa-6152d15afda0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.183769 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-scripts\") pod \"7fdf78a6-b107-4949-affa-6152d15afda0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.183799 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fdf78a6-b107-4949-affa-6152d15afda0-log-httpd\") pod \"7fdf78a6-b107-4949-affa-6152d15afda0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.183867 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk4fk\" (UniqueName: \"kubernetes.io/projected/7fdf78a6-b107-4949-affa-6152d15afda0-kube-api-access-lk4fk\") pod \"7fdf78a6-b107-4949-affa-6152d15afda0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.185522 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-sg-core-conf-yaml\") pod \"7fdf78a6-b107-4949-affa-6152d15afda0\" (UID: \"7fdf78a6-b107-4949-affa-6152d15afda0\") " Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.183903 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fdf78a6-b107-4949-affa-6152d15afda0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7fdf78a6-b107-4949-affa-6152d15afda0" (UID: "7fdf78a6-b107-4949-affa-6152d15afda0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.185422 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fdf78a6-b107-4949-affa-6152d15afda0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7fdf78a6-b107-4949-affa-6152d15afda0" (UID: "7fdf78a6-b107-4949-affa-6152d15afda0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.187004 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fdf78a6-b107-4949-affa-6152d15afda0-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.187098 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fdf78a6-b107-4949-affa-6152d15afda0-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.190156 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-scripts" (OuterVolumeSpecName: "scripts") pod "7fdf78a6-b107-4949-affa-6152d15afda0" (UID: "7fdf78a6-b107-4949-affa-6152d15afda0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.191746 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fdf78a6-b107-4949-affa-6152d15afda0-kube-api-access-lk4fk" (OuterVolumeSpecName: "kube-api-access-lk4fk") pod "7fdf78a6-b107-4949-affa-6152d15afda0" (UID: "7fdf78a6-b107-4949-affa-6152d15afda0"). InnerVolumeSpecName "kube-api-access-lk4fk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.222723 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7fdf78a6-b107-4949-affa-6152d15afda0" (UID: "7fdf78a6-b107-4949-affa-6152d15afda0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.286500 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7fdf78a6-b107-4949-affa-6152d15afda0" (UID: "7fdf78a6-b107-4949-affa-6152d15afda0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.288694 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.288725 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.288735 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk4fk\" (UniqueName: \"kubernetes.io/projected/7fdf78a6-b107-4949-affa-6152d15afda0-kube-api-access-lk4fk\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.288748 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.301845 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-config-data" (OuterVolumeSpecName: "config-data") pod "7fdf78a6-b107-4949-affa-6152d15afda0" (UID: "7fdf78a6-b107-4949-affa-6152d15afda0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.390099 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fdf78a6-b107-4949-affa-6152d15afda0-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.951413 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fdf78a6-b107-4949-affa-6152d15afda0","Type":"ContainerDied","Data":"0199a14e4f5a12a8019f727e78e7f15430a4ee84de969cda88096ca15bbed15b"} Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.951706 4775 scope.go:117] "RemoveContainer" containerID="ef9369d555365416b278955eec54a4b9a28ad97b7c0e8c0f39d301bd6d2b5a92" Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.951459 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.980646 4775 scope.go:117] "RemoveContainer" containerID="02056fc0b14ba46d658b8a5aac716fe485d874dd4d3feb41a0e56ea871cdd2a0" Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.992375 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:10:38 crc kubenswrapper[4775]: I0321 05:10:38.998095 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.004126 4775 scope.go:117] "RemoveContainer" containerID="5c6a6597ad7154fd7a253fa86eb6c5cde8565f20b6949e97002d664b75764604" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.025600 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:10:39 crc kubenswrapper[4775]: E0321 05:10:39.026088 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fdf78a6-b107-4949-affa-6152d15afda0" containerName="ceilometer-notification-agent" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.026109 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fdf78a6-b107-4949-affa-6152d15afda0" containerName="ceilometer-notification-agent" Mar 21 05:10:39 crc kubenswrapper[4775]: E0321 05:10:39.026145 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fdf78a6-b107-4949-affa-6152d15afda0" containerName="ceilometer-central-agent" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.026155 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fdf78a6-b107-4949-affa-6152d15afda0" containerName="ceilometer-central-agent" Mar 21 05:10:39 crc kubenswrapper[4775]: E0321 05:10:39.026186 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fdf78a6-b107-4949-affa-6152d15afda0" containerName="sg-core" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.026196 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fdf78a6-b107-4949-affa-6152d15afda0" containerName="sg-core" Mar 21 05:10:39 crc kubenswrapper[4775]: E0321 05:10:39.026212 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fdf78a6-b107-4949-affa-6152d15afda0" containerName="proxy-httpd" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.026219 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fdf78a6-b107-4949-affa-6152d15afda0" containerName="proxy-httpd" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.026440 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fdf78a6-b107-4949-affa-6152d15afda0" containerName="proxy-httpd" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.026467 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fdf78a6-b107-4949-affa-6152d15afda0" containerName="ceilometer-central-agent" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.026487 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fdf78a6-b107-4949-affa-6152d15afda0" containerName="sg-core" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.026507 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fdf78a6-b107-4949-affa-6152d15afda0" containerName="ceilometer-notification-agent" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.028697 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.030420 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.032379 4775 scope.go:117] "RemoveContainer" containerID="8f8cabcb03903e4d89761a18703dd5c97870cd2e3d0ef375a4910ff30a8d99cf" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.032661 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.033051 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.051518 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.104778 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15d97495-428d-47e0-a115-99c7fd08850a-scripts\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.104867 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59bs7\" (UniqueName: \"kubernetes.io/projected/15d97495-428d-47e0-a115-99c7fd08850a-kube-api-access-59bs7\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.104889 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15d97495-428d-47e0-a115-99c7fd08850a-run-httpd\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.104911 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15d97495-428d-47e0-a115-99c7fd08850a-log-httpd\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.105179 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/15d97495-428d-47e0-a115-99c7fd08850a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.105298 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15d97495-428d-47e0-a115-99c7fd08850a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.105379 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15d97495-428d-47e0-a115-99c7fd08850a-config-data\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.105406 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/15d97495-428d-47e0-a115-99c7fd08850a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.206920 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15d97495-428d-47e0-a115-99c7fd08850a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.206981 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15d97495-428d-47e0-a115-99c7fd08850a-config-data\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.207003 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/15d97495-428d-47e0-a115-99c7fd08850a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.207042 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15d97495-428d-47e0-a115-99c7fd08850a-scripts\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.207106 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59bs7\" (UniqueName: \"kubernetes.io/projected/15d97495-428d-47e0-a115-99c7fd08850a-kube-api-access-59bs7\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.207153 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15d97495-428d-47e0-a115-99c7fd08850a-run-httpd\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.207172 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15d97495-428d-47e0-a115-99c7fd08850a-log-httpd\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.207253 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/15d97495-428d-47e0-a115-99c7fd08850a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.208989 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15d97495-428d-47e0-a115-99c7fd08850a-run-httpd\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.208994 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15d97495-428d-47e0-a115-99c7fd08850a-log-httpd\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.213847 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15d97495-428d-47e0-a115-99c7fd08850a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.214225 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/15d97495-428d-47e0-a115-99c7fd08850a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.220635 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15d97495-428d-47e0-a115-99c7fd08850a-config-data\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.226547 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/15d97495-428d-47e0-a115-99c7fd08850a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.228723 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15d97495-428d-47e0-a115-99c7fd08850a-scripts\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.230625 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59bs7\" (UniqueName: \"kubernetes.io/projected/15d97495-428d-47e0-a115-99c7fd08850a-kube-api-access-59bs7\") pod \"ceilometer-0\" (UID: \"15d97495-428d-47e0-a115-99c7fd08850a\") " pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.358657 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.678897 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fdf78a6-b107-4949-affa-6152d15afda0" path="/var/lib/kubelet/pods/7fdf78a6-b107-4949-affa-6152d15afda0/volumes" Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.813844 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 05:10:39 crc kubenswrapper[4775]: W0321 05:10:39.816946 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15d97495_428d_47e0_a115_99c7fd08850a.slice/crio-1622581afed8ba9cf1c1c7cb429c446134537918cd31333b20a211b831399aa7 WatchSource:0}: Error finding container 1622581afed8ba9cf1c1c7cb429c446134537918cd31333b20a211b831399aa7: Status 404 returned error can't find the container with id 1622581afed8ba9cf1c1c7cb429c446134537918cd31333b20a211b831399aa7 Mar 21 05:10:39 crc kubenswrapper[4775]: I0321 05:10:39.963272 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15d97495-428d-47e0-a115-99c7fd08850a","Type":"ContainerStarted","Data":"1622581afed8ba9cf1c1c7cb429c446134537918cd31333b20a211b831399aa7"} Mar 21 05:10:41 crc kubenswrapper[4775]: I0321 05:10:41.984006 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15d97495-428d-47e0-a115-99c7fd08850a","Type":"ContainerStarted","Data":"24e9203926ce20d73abad431ad234bd89366d9c12b6fe8212d5a1ee4d29a327c"} Mar 21 05:10:41 crc kubenswrapper[4775]: I0321 05:10:41.984602 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15d97495-428d-47e0-a115-99c7fd08850a","Type":"ContainerStarted","Data":"4d7ec2197c044f2c92cf604a4c8062ca6bb12683a9edaf586742518fe3f3dabf"} Mar 21 05:10:42 crc kubenswrapper[4775]: I0321 05:10:42.139396 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 21 05:10:42 crc kubenswrapper[4775]: I0321 05:10:42.573862 4775 scope.go:117] "RemoveContainer" containerID="ca9c08541581bc07e396b682f2ba269a753d458f0a5cc3ff5f26ace1828dea81" Mar 21 05:10:42 crc kubenswrapper[4775]: I0321 05:10:42.995926 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15d97495-428d-47e0-a115-99c7fd08850a","Type":"ContainerStarted","Data":"2ddaecf980b5884d76751d1c960d095146d81d5cd9ff2519987921267a5e42d3"} Mar 21 05:10:43 crc kubenswrapper[4775]: I0321 05:10:43.248325 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 21 05:10:45 crc kubenswrapper[4775]: I0321 05:10:45.028454 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15d97495-428d-47e0-a115-99c7fd08850a","Type":"ContainerStarted","Data":"3deed700d08f13985c8e9b21af8cafbca8a7db18fd1bc5bc6c6276a4d26500ae"} Mar 21 05:10:45 crc kubenswrapper[4775]: I0321 05:10:45.030102 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 21 05:10:45 crc kubenswrapper[4775]: I0321 05:10:45.066455 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.876802602 podStartE2EDuration="7.066433712s" podCreationTimestamp="2026-03-21 05:10:38 +0000 UTC" firstStartedPulling="2026-03-21 05:10:39.820386857 +0000 UTC m=+1392.796850481" lastFinishedPulling="2026-03-21 05:10:44.010017967 +0000 UTC m=+1396.986481591" observedRunningTime="2026-03-21 05:10:45.056715247 +0000 UTC m=+1398.033178881" watchObservedRunningTime="2026-03-21 05:10:45.066433712 +0000 UTC m=+1398.042897336" Mar 21 05:10:46 crc kubenswrapper[4775]: I0321 05:10:46.296575 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 21 05:10:46 crc kubenswrapper[4775]: I0321 05:10:46.741364 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="839e915e-8197-48e9-8b69-56ac420a1eed" containerName="rabbitmq" containerID="cri-o://c1cc3f5c012ce3f528f11e083192eb0842f7dad8501df783855a594cab5449da" gracePeriod=604796 Mar 21 05:10:47 crc kubenswrapper[4775]: I0321 05:10:47.742009 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="375fb8b7-b673-4fd7-ae51-5f82f33c196f" containerName="rabbitmq" containerID="cri-o://2b205fc4ed697ad5cbf0298ac1f6315a11af8ff066df4aa11b086ace7ab697d3" gracePeriod=604796 Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.100552 4775 generic.go:334] "Generic (PLEG): container finished" podID="839e915e-8197-48e9-8b69-56ac420a1eed" containerID="c1cc3f5c012ce3f528f11e083192eb0842f7dad8501df783855a594cab5449da" exitCode=0 Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.100571 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"839e915e-8197-48e9-8b69-56ac420a1eed","Type":"ContainerDied","Data":"c1cc3f5c012ce3f528f11e083192eb0842f7dad8501df783855a594cab5449da"} Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.356065 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.377428 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-plugins\") pod \"839e915e-8197-48e9-8b69-56ac420a1eed\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.377477 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/839e915e-8197-48e9-8b69-56ac420a1eed-plugins-conf\") pod \"839e915e-8197-48e9-8b69-56ac420a1eed\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.377554 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/839e915e-8197-48e9-8b69-56ac420a1eed-server-conf\") pod \"839e915e-8197-48e9-8b69-56ac420a1eed\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.378234 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmtzj\" (UniqueName: \"kubernetes.io/projected/839e915e-8197-48e9-8b69-56ac420a1eed-kube-api-access-lmtzj\") pod \"839e915e-8197-48e9-8b69-56ac420a1eed\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.378327 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-erlang-cookie\") pod \"839e915e-8197-48e9-8b69-56ac420a1eed\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.378453 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-confd\") pod \"839e915e-8197-48e9-8b69-56ac420a1eed\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.378498 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/839e915e-8197-48e9-8b69-56ac420a1eed-config-data\") pod \"839e915e-8197-48e9-8b69-56ac420a1eed\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.378790 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "839e915e-8197-48e9-8b69-56ac420a1eed" (UID: "839e915e-8197-48e9-8b69-56ac420a1eed"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.379016 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/839e915e-8197-48e9-8b69-56ac420a1eed-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "839e915e-8197-48e9-8b69-56ac420a1eed" (UID: "839e915e-8197-48e9-8b69-56ac420a1eed"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.379277 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "839e915e-8197-48e9-8b69-56ac420a1eed" (UID: "839e915e-8197-48e9-8b69-56ac420a1eed"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.385027 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/839e915e-8197-48e9-8b69-56ac420a1eed-kube-api-access-lmtzj" (OuterVolumeSpecName: "kube-api-access-lmtzj") pod "839e915e-8197-48e9-8b69-56ac420a1eed" (UID: "839e915e-8197-48e9-8b69-56ac420a1eed"). InnerVolumeSpecName "kube-api-access-lmtzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.392572 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"839e915e-8197-48e9-8b69-56ac420a1eed\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.392670 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/839e915e-8197-48e9-8b69-56ac420a1eed-erlang-cookie-secret\") pod \"839e915e-8197-48e9-8b69-56ac420a1eed\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.392715 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/839e915e-8197-48e9-8b69-56ac420a1eed-pod-info\") pod \"839e915e-8197-48e9-8b69-56ac420a1eed\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.392777 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-tls\") pod \"839e915e-8197-48e9-8b69-56ac420a1eed\" (UID: \"839e915e-8197-48e9-8b69-56ac420a1eed\") " Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.395837 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.395910 4775 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/839e915e-8197-48e9-8b69-56ac420a1eed-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.395926 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmtzj\" (UniqueName: \"kubernetes.io/projected/839e915e-8197-48e9-8b69-56ac420a1eed-kube-api-access-lmtzj\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.395944 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.403591 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/839e915e-8197-48e9-8b69-56ac420a1eed-pod-info" (OuterVolumeSpecName: "pod-info") pod "839e915e-8197-48e9-8b69-56ac420a1eed" (UID: "839e915e-8197-48e9-8b69-56ac420a1eed"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.403743 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "839e915e-8197-48e9-8b69-56ac420a1eed" (UID: "839e915e-8197-48e9-8b69-56ac420a1eed"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.412296 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/839e915e-8197-48e9-8b69-56ac420a1eed-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "839e915e-8197-48e9-8b69-56ac420a1eed" (UID: "839e915e-8197-48e9-8b69-56ac420a1eed"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.412594 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "839e915e-8197-48e9-8b69-56ac420a1eed" (UID: "839e915e-8197-48e9-8b69-56ac420a1eed"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.446457 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/839e915e-8197-48e9-8b69-56ac420a1eed-config-data" (OuterVolumeSpecName: "config-data") pod "839e915e-8197-48e9-8b69-56ac420a1eed" (UID: "839e915e-8197-48e9-8b69-56ac420a1eed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.487234 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/839e915e-8197-48e9-8b69-56ac420a1eed-server-conf" (OuterVolumeSpecName: "server-conf") pod "839e915e-8197-48e9-8b69-56ac420a1eed" (UID: "839e915e-8197-48e9-8b69-56ac420a1eed"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.497580 4775 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/839e915e-8197-48e9-8b69-56ac420a1eed-server-conf\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.497616 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/839e915e-8197-48e9-8b69-56ac420a1eed-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.497645 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.497658 4775 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/839e915e-8197-48e9-8b69-56ac420a1eed-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.497669 4775 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/839e915e-8197-48e9-8b69-56ac420a1eed-pod-info\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.497679 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.584334 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.586291 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "839e915e-8197-48e9-8b69-56ac420a1eed" (UID: "839e915e-8197-48e9-8b69-56ac420a1eed"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.614476 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:53 crc kubenswrapper[4775]: I0321 05:10:53.614781 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/839e915e-8197-48e9-8b69-56ac420a1eed-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.116466 4775 generic.go:334] "Generic (PLEG): container finished" podID="375fb8b7-b673-4fd7-ae51-5f82f33c196f" containerID="2b205fc4ed697ad5cbf0298ac1f6315a11af8ff066df4aa11b086ace7ab697d3" exitCode=0 Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.116572 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"375fb8b7-b673-4fd7-ae51-5f82f33c196f","Type":"ContainerDied","Data":"2b205fc4ed697ad5cbf0298ac1f6315a11af8ff066df4aa11b086ace7ab697d3"} Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.120412 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"839e915e-8197-48e9-8b69-56ac420a1eed","Type":"ContainerDied","Data":"ef63fbfadda52109f2b3ecef0f45ed06c3a90225a2c4d5559784277a4c40c9dd"} Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.120478 4775 scope.go:117] "RemoveContainer" containerID="c1cc3f5c012ce3f528f11e083192eb0842f7dad8501df783855a594cab5449da" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.120483 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.159710 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.175212 4775 scope.go:117] "RemoveContainer" containerID="5a7046cba7a8f5bae4cf5ebd270104b81a62d747ce978d3680ad2d2e0a16d243" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.175366 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.231247 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 21 05:10:54 crc kubenswrapper[4775]: E0321 05:10:54.231713 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="839e915e-8197-48e9-8b69-56ac420a1eed" containerName="rabbitmq" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.231726 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="839e915e-8197-48e9-8b69-56ac420a1eed" containerName="rabbitmq" Mar 21 05:10:54 crc kubenswrapper[4775]: E0321 05:10:54.231742 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="839e915e-8197-48e9-8b69-56ac420a1eed" containerName="setup-container" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.231751 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="839e915e-8197-48e9-8b69-56ac420a1eed" containerName="setup-container" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.231961 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="839e915e-8197-48e9-8b69-56ac420a1eed" containerName="rabbitmq" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.232901 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.235661 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.235879 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.236196 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.236305 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.236466 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.236977 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-kmjnx" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.237136 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.246158 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.341040 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8t8r\" (UniqueName: \"kubernetes.io/projected/e5e83941-a38d-4ee9-b967-1dac69c5a55b-kube-api-access-c8t8r\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.341082 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e5e83941-a38d-4ee9-b967-1dac69c5a55b-config-data\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.341139 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e5e83941-a38d-4ee9-b967-1dac69c5a55b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.341408 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e5e83941-a38d-4ee9-b967-1dac69c5a55b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.341451 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e5e83941-a38d-4ee9-b967-1dac69c5a55b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.341551 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e5e83941-a38d-4ee9-b967-1dac69c5a55b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.341773 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e5e83941-a38d-4ee9-b967-1dac69c5a55b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.341826 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e5e83941-a38d-4ee9-b967-1dac69c5a55b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.341935 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.341973 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e5e83941-a38d-4ee9-b967-1dac69c5a55b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.341997 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e5e83941-a38d-4ee9-b967-1dac69c5a55b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.398182 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.443113 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-erlang-cookie\") pod \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.443195 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-plugins\") pod \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.443244 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.443364 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/375fb8b7-b673-4fd7-ae51-5f82f33c196f-pod-info\") pod \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.443394 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/375fb8b7-b673-4fd7-ae51-5f82f33c196f-config-data\") pod \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.443427 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-tls\") pod \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.443452 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-confd\") pod \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.443514 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqdt6\" (UniqueName: \"kubernetes.io/projected/375fb8b7-b673-4fd7-ae51-5f82f33c196f-kube-api-access-vqdt6\") pod \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.443558 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/375fb8b7-b673-4fd7-ae51-5f82f33c196f-plugins-conf\") pod \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.443580 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/375fb8b7-b673-4fd7-ae51-5f82f33c196f-erlang-cookie-secret\") pod \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.443598 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/375fb8b7-b673-4fd7-ae51-5f82f33c196f-server-conf\") pod \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\" (UID: \"375fb8b7-b673-4fd7-ae51-5f82f33c196f\") " Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.443888 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.443924 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e5e83941-a38d-4ee9-b967-1dac69c5a55b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.443966 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e5e83941-a38d-4ee9-b967-1dac69c5a55b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.444020 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8t8r\" (UniqueName: \"kubernetes.io/projected/e5e83941-a38d-4ee9-b967-1dac69c5a55b-kube-api-access-c8t8r\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.444051 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e5e83941-a38d-4ee9-b967-1dac69c5a55b-config-data\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.444094 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e5e83941-a38d-4ee9-b967-1dac69c5a55b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.444229 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e5e83941-a38d-4ee9-b967-1dac69c5a55b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.444251 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e5e83941-a38d-4ee9-b967-1dac69c5a55b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.444270 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e5e83941-a38d-4ee9-b967-1dac69c5a55b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.444387 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e5e83941-a38d-4ee9-b967-1dac69c5a55b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.444416 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e5e83941-a38d-4ee9-b967-1dac69c5a55b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.445774 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e5e83941-a38d-4ee9-b967-1dac69c5a55b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.446154 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e5e83941-a38d-4ee9-b967-1dac69c5a55b-config-data\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.446615 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e5e83941-a38d-4ee9-b967-1dac69c5a55b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.446841 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.447079 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e5e83941-a38d-4ee9-b967-1dac69c5a55b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.447805 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "375fb8b7-b673-4fd7-ae51-5f82f33c196f" (UID: "375fb8b7-b673-4fd7-ae51-5f82f33c196f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.447909 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e5e83941-a38d-4ee9-b967-1dac69c5a55b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.448702 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/375fb8b7-b673-4fd7-ae51-5f82f33c196f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "375fb8b7-b673-4fd7-ae51-5f82f33c196f" (UID: "375fb8b7-b673-4fd7-ae51-5f82f33c196f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.450149 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "375fb8b7-b673-4fd7-ae51-5f82f33c196f" (UID: "375fb8b7-b673-4fd7-ae51-5f82f33c196f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.454885 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e5e83941-a38d-4ee9-b967-1dac69c5a55b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.458667 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e5e83941-a38d-4ee9-b967-1dac69c5a55b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.460890 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/375fb8b7-b673-4fd7-ae51-5f82f33c196f-kube-api-access-vqdt6" (OuterVolumeSpecName: "kube-api-access-vqdt6") pod "375fb8b7-b673-4fd7-ae51-5f82f33c196f" (UID: "375fb8b7-b673-4fd7-ae51-5f82f33c196f"). InnerVolumeSpecName "kube-api-access-vqdt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.462108 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e5e83941-a38d-4ee9-b967-1dac69c5a55b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.462132 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "375fb8b7-b673-4fd7-ae51-5f82f33c196f" (UID: "375fb8b7-b673-4fd7-ae51-5f82f33c196f"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.463068 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/375fb8b7-b673-4fd7-ae51-5f82f33c196f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "375fb8b7-b673-4fd7-ae51-5f82f33c196f" (UID: "375fb8b7-b673-4fd7-ae51-5f82f33c196f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.473300 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "375fb8b7-b673-4fd7-ae51-5f82f33c196f" (UID: "375fb8b7-b673-4fd7-ae51-5f82f33c196f"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.473550 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8t8r\" (UniqueName: \"kubernetes.io/projected/e5e83941-a38d-4ee9-b967-1dac69c5a55b-kube-api-access-c8t8r\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.473938 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e5e83941-a38d-4ee9-b967-1dac69c5a55b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.474790 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/375fb8b7-b673-4fd7-ae51-5f82f33c196f-pod-info" (OuterVolumeSpecName: "pod-info") pod "375fb8b7-b673-4fd7-ae51-5f82f33c196f" (UID: "375fb8b7-b673-4fd7-ae51-5f82f33c196f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.517233 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/375fb8b7-b673-4fd7-ae51-5f82f33c196f-config-data" (OuterVolumeSpecName: "config-data") pod "375fb8b7-b673-4fd7-ae51-5f82f33c196f" (UID: "375fb8b7-b673-4fd7-ae51-5f82f33c196f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.536981 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"e5e83941-a38d-4ee9-b967-1dac69c5a55b\") " pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.546068 4775 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/375fb8b7-b673-4fd7-ae51-5f82f33c196f-pod-info\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.546109 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/375fb8b7-b673-4fd7-ae51-5f82f33c196f-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.546137 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.546149 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqdt6\" (UniqueName: \"kubernetes.io/projected/375fb8b7-b673-4fd7-ae51-5f82f33c196f-kube-api-access-vqdt6\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.546161 4775 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/375fb8b7-b673-4fd7-ae51-5f82f33c196f-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.546172 4775 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/375fb8b7-b673-4fd7-ae51-5f82f33c196f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.546183 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.546194 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.546221 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.560496 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/375fb8b7-b673-4fd7-ae51-5f82f33c196f-server-conf" (OuterVolumeSpecName: "server-conf") pod "375fb8b7-b673-4fd7-ae51-5f82f33c196f" (UID: "375fb8b7-b673-4fd7-ae51-5f82f33c196f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.571854 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.582177 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.594074 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "375fb8b7-b673-4fd7-ae51-5f82f33c196f" (UID: "375fb8b7-b673-4fd7-ae51-5f82f33c196f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.647609 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.647642 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/375fb8b7-b673-4fd7-ae51-5f82f33c196f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:54 crc kubenswrapper[4775]: I0321 05:10:54.647653 4775 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/375fb8b7-b673-4fd7-ae51-5f82f33c196f-server-conf\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.004810 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d558885bc-9m26s"] Mar 21 05:10:55 crc kubenswrapper[4775]: E0321 05:10:55.005576 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="375fb8b7-b673-4fd7-ae51-5f82f33c196f" containerName="rabbitmq" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.005596 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="375fb8b7-b673-4fd7-ae51-5f82f33c196f" containerName="rabbitmq" Mar 21 05:10:55 crc kubenswrapper[4775]: E0321 05:10:55.005613 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="375fb8b7-b673-4fd7-ae51-5f82f33c196f" containerName="setup-container" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.005622 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="375fb8b7-b673-4fd7-ae51-5f82f33c196f" containerName="setup-container" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.005848 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="375fb8b7-b673-4fd7-ae51-5f82f33c196f" containerName="rabbitmq" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.007012 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.011018 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.019591 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-9m26s"] Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.056633 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.056711 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.056737 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq54r\" (UniqueName: \"kubernetes.io/projected/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-kube-api-access-nq54r\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.056876 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.056927 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-dns-svc\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.056947 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-config\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.056983 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.078249 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.150976 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e5e83941-a38d-4ee9-b967-1dac69c5a55b","Type":"ContainerStarted","Data":"9f1c3a538f47d58bc76013b107bdff68081d9c8cfd6e7ffdae9680d305e10223"} Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.156106 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"375fb8b7-b673-4fd7-ae51-5f82f33c196f","Type":"ContainerDied","Data":"56ec7d33ff14d09de4f061f404a6116851871ee2ca592e6f5daea8ccedd576e8"} Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.156170 4775 scope.go:117] "RemoveContainer" containerID="2b205fc4ed697ad5cbf0298ac1f6315a11af8ff066df4aa11b086ace7ab697d3" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.156207 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.158318 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.158366 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-dns-svc\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.158384 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-config\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.158413 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.158457 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.158481 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.158495 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq54r\" (UniqueName: \"kubernetes.io/projected/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-kube-api-access-nq54r\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.159582 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.160090 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.160940 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.162533 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-config\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.163200 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.163860 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-dns-svc\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.164286 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-9m26s"] Mar 21 05:10:55 crc kubenswrapper[4775]: E0321 05:10:55.167599 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-nq54r], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-d558885bc-9m26s" podUID="eb994900-6140-4a10-aafd-0e9a6d6dfcc9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.186721 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq54r\" (UniqueName: \"kubernetes.io/projected/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-kube-api-access-nq54r\") pod \"dnsmasq-dns-d558885bc-9m26s\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.203672 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-mmgp9"] Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.205549 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.226786 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-mmgp9"] Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.250690 4775 scope.go:117] "RemoveContainer" containerID="1c4292485e2c0ee4f0c83f962fffc48a32f990f8105220ac94e44a4691b7ff1f" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.259766 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4918607c-6074-4fb3-a0a0-8def479058a0-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.259827 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4918607c-6074-4fb3-a0a0-8def479058a0-config\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.259886 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghvj6\" (UniqueName: \"kubernetes.io/projected/4918607c-6074-4fb3-a0a0-8def479058a0-kube-api-access-ghvj6\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.259905 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4918607c-6074-4fb3-a0a0-8def479058a0-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.259924 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4918607c-6074-4fb3-a0a0-8def479058a0-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.259959 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4918607c-6074-4fb3-a0a0-8def479058a0-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.260037 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4918607c-6074-4fb3-a0a0-8def479058a0-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.295359 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.311160 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.338332 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.343658 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.346504 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.346980 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.347848 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.347934 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.348038 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.348180 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.348343 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-5pn7h" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.356084 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.426296 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4918607c-6074-4fb3-a0a0-8def479058a0-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.426381 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c95486b5-f2ad-4098-912d-6749b329824b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.426483 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c95486b5-f2ad-4098-912d-6749b329824b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.426661 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c95486b5-f2ad-4098-912d-6749b329824b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.426724 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4918607c-6074-4fb3-a0a0-8def479058a0-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.426774 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dn4q\" (UniqueName: \"kubernetes.io/projected/c95486b5-f2ad-4098-912d-6749b329824b-kube-api-access-2dn4q\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.426805 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c95486b5-f2ad-4098-912d-6749b329824b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.426907 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4918607c-6074-4fb3-a0a0-8def479058a0-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.426960 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c95486b5-f2ad-4098-912d-6749b329824b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.427012 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c95486b5-f2ad-4098-912d-6749b329824b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.427044 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.427080 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4918607c-6074-4fb3-a0a0-8def479058a0-config\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.427103 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c95486b5-f2ad-4098-912d-6749b329824b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.427180 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c95486b5-f2ad-4098-912d-6749b329824b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.427245 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghvj6\" (UniqueName: \"kubernetes.io/projected/4918607c-6074-4fb3-a0a0-8def479058a0-kube-api-access-ghvj6\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.427302 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c95486b5-f2ad-4098-912d-6749b329824b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.427338 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4918607c-6074-4fb3-a0a0-8def479058a0-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.427368 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4918607c-6074-4fb3-a0a0-8def479058a0-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.428384 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4918607c-6074-4fb3-a0a0-8def479058a0-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.428721 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4918607c-6074-4fb3-a0a0-8def479058a0-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.429021 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4918607c-6074-4fb3-a0a0-8def479058a0-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.430920 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4918607c-6074-4fb3-a0a0-8def479058a0-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.431050 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4918607c-6074-4fb3-a0a0-8def479058a0-config\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.431211 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4918607c-6074-4fb3-a0a0-8def479058a0-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.445448 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghvj6\" (UniqueName: \"kubernetes.io/projected/4918607c-6074-4fb3-a0a0-8def479058a0-kube-api-access-ghvj6\") pod \"dnsmasq-dns-78c64bc9c5-mmgp9\" (UID: \"4918607c-6074-4fb3-a0a0-8def479058a0\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.528569 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c95486b5-f2ad-4098-912d-6749b329824b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.528619 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c95486b5-f2ad-4098-912d-6749b329824b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.528652 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dn4q\" (UniqueName: \"kubernetes.io/projected/c95486b5-f2ad-4098-912d-6749b329824b-kube-api-access-2dn4q\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.528670 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c95486b5-f2ad-4098-912d-6749b329824b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.528710 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c95486b5-f2ad-4098-912d-6749b329824b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.528729 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c95486b5-f2ad-4098-912d-6749b329824b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.528749 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.528768 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c95486b5-f2ad-4098-912d-6749b329824b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.528793 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c95486b5-f2ad-4098-912d-6749b329824b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.528820 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c95486b5-f2ad-4098-912d-6749b329824b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.528869 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c95486b5-f2ad-4098-912d-6749b329824b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.530174 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c95486b5-f2ad-4098-912d-6749b329824b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.530387 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.530688 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c95486b5-f2ad-4098-912d-6749b329824b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.530726 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c95486b5-f2ad-4098-912d-6749b329824b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.530876 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c95486b5-f2ad-4098-912d-6749b329824b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.532087 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c95486b5-f2ad-4098-912d-6749b329824b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.532450 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c95486b5-f2ad-4098-912d-6749b329824b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.534280 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c95486b5-f2ad-4098-912d-6749b329824b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.534925 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c95486b5-f2ad-4098-912d-6749b329824b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.537839 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c95486b5-f2ad-4098-912d-6749b329824b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.549231 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dn4q\" (UniqueName: \"kubernetes.io/projected/c95486b5-f2ad-4098-912d-6749b329824b-kube-api-access-2dn4q\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.571202 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c95486b5-f2ad-4098-912d-6749b329824b\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.590356 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.664243 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.695333 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="375fb8b7-b673-4fd7-ae51-5f82f33c196f" path="/var/lib/kubelet/pods/375fb8b7-b673-4fd7-ae51-5f82f33c196f/volumes" Mar 21 05:10:55 crc kubenswrapper[4775]: I0321 05:10:55.696516 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="839e915e-8197-48e9-8b69-56ac420a1eed" path="/var/lib/kubelet/pods/839e915e-8197-48e9-8b69-56ac420a1eed/volumes" Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.058499 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-mmgp9"] Mar 21 05:10:56 crc kubenswrapper[4775]: W0321 05:10:56.065081 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4918607c_6074_4fb3_a0a0_8def479058a0.slice/crio-062d775e8dc7fe83d1d19047e5f49f7fb2f5f14974bc435df1465a3936dcebc8 WatchSource:0}: Error finding container 062d775e8dc7fe83d1d19047e5f49f7fb2f5f14974bc435df1465a3936dcebc8: Status 404 returned error can't find the container with id 062d775e8dc7fe83d1d19047e5f49f7fb2f5f14974bc435df1465a3936dcebc8 Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.170813 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" event={"ID":"4918607c-6074-4fb3-a0a0-8def479058a0","Type":"ContainerStarted","Data":"062d775e8dc7fe83d1d19047e5f49f7fb2f5f14974bc435df1465a3936dcebc8"} Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.170833 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.188498 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 21 05:10:56 crc kubenswrapper[4775]: W0321 05:10:56.190255 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc95486b5_f2ad_4098_912d_6749b329824b.slice/crio-1bb5e8dd2af95a8994651214c672edc0ac0ebfc669a4e75bf466f382e4a51488 WatchSource:0}: Error finding container 1bb5e8dd2af95a8994651214c672edc0ac0ebfc669a4e75bf466f382e4a51488: Status 404 returned error can't find the container with id 1bb5e8dd2af95a8994651214c672edc0ac0ebfc669a4e75bf466f382e4a51488 Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.324452 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.444611 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nq54r\" (UniqueName: \"kubernetes.io/projected/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-kube-api-access-nq54r\") pod \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.444690 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-openstack-edpm-ipam\") pod \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.444760 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-dns-svc\") pod \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.444796 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-ovsdbserver-sb\") pod \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.444854 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-dns-swift-storage-0\") pod \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.444975 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-ovsdbserver-nb\") pod \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.445049 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-config\") pod \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\" (UID: \"eb994900-6140-4a10-aafd-0e9a6d6dfcc9\") " Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.445918 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-config" (OuterVolumeSpecName: "config") pod "eb994900-6140-4a10-aafd-0e9a6d6dfcc9" (UID: "eb994900-6140-4a10-aafd-0e9a6d6dfcc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.446471 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "eb994900-6140-4a10-aafd-0e9a6d6dfcc9" (UID: "eb994900-6140-4a10-aafd-0e9a6d6dfcc9"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.446800 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eb994900-6140-4a10-aafd-0e9a6d6dfcc9" (UID: "eb994900-6140-4a10-aafd-0e9a6d6dfcc9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.447149 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eb994900-6140-4a10-aafd-0e9a6d6dfcc9" (UID: "eb994900-6140-4a10-aafd-0e9a6d6dfcc9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.447523 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "eb994900-6140-4a10-aafd-0e9a6d6dfcc9" (UID: "eb994900-6140-4a10-aafd-0e9a6d6dfcc9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.447897 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eb994900-6140-4a10-aafd-0e9a6d6dfcc9" (UID: "eb994900-6140-4a10-aafd-0e9a6d6dfcc9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.450486 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-kube-api-access-nq54r" (OuterVolumeSpecName: "kube-api-access-nq54r") pod "eb994900-6140-4a10-aafd-0e9a6d6dfcc9" (UID: "eb994900-6140-4a10-aafd-0e9a6d6dfcc9"). InnerVolumeSpecName "kube-api-access-nq54r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.547224 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.547262 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.547272 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.547280 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nq54r\" (UniqueName: \"kubernetes.io/projected/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-kube-api-access-nq54r\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.547289 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.547296 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:56 crc kubenswrapper[4775]: I0321 05:10:56.547305 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb994900-6140-4a10-aafd-0e9a6d6dfcc9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 05:10:57 crc kubenswrapper[4775]: I0321 05:10:57.184206 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e5e83941-a38d-4ee9-b967-1dac69c5a55b","Type":"ContainerStarted","Data":"92cf1e237e929c497cbf8f575434685971036de8960a393e86900a1e6642f0ed"} Mar 21 05:10:57 crc kubenswrapper[4775]: I0321 05:10:57.187571 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c95486b5-f2ad-4098-912d-6749b329824b","Type":"ContainerStarted","Data":"1bb5e8dd2af95a8994651214c672edc0ac0ebfc669a4e75bf466f382e4a51488"} Mar 21 05:10:57 crc kubenswrapper[4775]: I0321 05:10:57.189635 4775 generic.go:334] "Generic (PLEG): container finished" podID="4918607c-6074-4fb3-a0a0-8def479058a0" containerID="83fa5958fe9b584a13b72951e70ec159f4909e5520901e4dceebd69c63cb7ca9" exitCode=0 Mar 21 05:10:57 crc kubenswrapper[4775]: I0321 05:10:57.189690 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-9m26s" Mar 21 05:10:57 crc kubenswrapper[4775]: I0321 05:10:57.189703 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" event={"ID":"4918607c-6074-4fb3-a0a0-8def479058a0","Type":"ContainerDied","Data":"83fa5958fe9b584a13b72951e70ec159f4909e5520901e4dceebd69c63cb7ca9"} Mar 21 05:10:57 crc kubenswrapper[4775]: I0321 05:10:57.420790 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-9m26s"] Mar 21 05:10:57 crc kubenswrapper[4775]: I0321 05:10:57.431392 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-9m26s"] Mar 21 05:10:57 crc kubenswrapper[4775]: I0321 05:10:57.680084 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb994900-6140-4a10-aafd-0e9a6d6dfcc9" path="/var/lib/kubelet/pods/eb994900-6140-4a10-aafd-0e9a6d6dfcc9/volumes" Mar 21 05:10:58 crc kubenswrapper[4775]: I0321 05:10:58.208360 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c95486b5-f2ad-4098-912d-6749b329824b","Type":"ContainerStarted","Data":"f565924e64e065b4ec412c0136fb28d6de7cffeff14b255d7fd47acd3851a689"} Mar 21 05:10:58 crc kubenswrapper[4775]: I0321 05:10:58.213284 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" event={"ID":"4918607c-6074-4fb3-a0a0-8def479058a0","Type":"ContainerStarted","Data":"32383fa63f060e6b7e3508ff1589d2a5f19f1eb11d6ebff99c27ff830eb0347f"} Mar 21 05:10:58 crc kubenswrapper[4775]: I0321 05:10:58.213492 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:10:58 crc kubenswrapper[4775]: I0321 05:10:58.249899 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" podStartSLOduration=3.249882356 podStartE2EDuration="3.249882356s" podCreationTimestamp="2026-03-21 05:10:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:10:58.24932509 +0000 UTC m=+1411.225788714" watchObservedRunningTime="2026-03-21 05:10:58.249882356 +0000 UTC m=+1411.226345980" Mar 21 05:11:02 crc kubenswrapper[4775]: I0321 05:11:02.482234 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:11:02 crc kubenswrapper[4775]: I0321 05:11:02.482525 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:11:02 crc kubenswrapper[4775]: I0321 05:11:02.482578 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 05:11:02 crc kubenswrapper[4775]: I0321 05:11:02.483382 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b6e85a7c4acc97394b06df812a396284f471ec9c7f9eee22918e9da54e21feda"} pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 05:11:02 crc kubenswrapper[4775]: I0321 05:11:02.483454 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" containerID="cri-o://b6e85a7c4acc97394b06df812a396284f471ec9c7f9eee22918e9da54e21feda" gracePeriod=600 Mar 21 05:11:03 crc kubenswrapper[4775]: I0321 05:11:03.256612 4775 generic.go:334] "Generic (PLEG): container finished" podID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerID="b6e85a7c4acc97394b06df812a396284f471ec9c7f9eee22918e9da54e21feda" exitCode=0 Mar 21 05:11:03 crc kubenswrapper[4775]: I0321 05:11:03.256652 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerDied","Data":"b6e85a7c4acc97394b06df812a396284f471ec9c7f9eee22918e9da54e21feda"} Mar 21 05:11:03 crc kubenswrapper[4775]: I0321 05:11:03.256937 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerStarted","Data":"5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4"} Mar 21 05:11:03 crc kubenswrapper[4775]: I0321 05:11:03.256959 4775 scope.go:117] "RemoveContainer" containerID="09c71a6e96dc622a58adf5b83a67eab26ff45301d2bfcc2a43f1cbd9eb2d9791" Mar 21 05:11:05 crc kubenswrapper[4775]: I0321 05:11:05.593235 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78c64bc9c5-mmgp9" Mar 21 05:11:05 crc kubenswrapper[4775]: I0321 05:11:05.691682 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-rtnbf"] Mar 21 05:11:05 crc kubenswrapper[4775]: I0321 05:11:05.691945 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" podUID="d725aabc-ba32-4c0e-bc91-8819d73cae40" containerName="dnsmasq-dns" containerID="cri-o://90f0ef44d50d3f68a6e80b10a816b133ae6b897ca8881964ca82e79aca5bf325" gracePeriod=10 Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.245910 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.289076 4775 generic.go:334] "Generic (PLEG): container finished" podID="d725aabc-ba32-4c0e-bc91-8819d73cae40" containerID="90f0ef44d50d3f68a6e80b10a816b133ae6b897ca8881964ca82e79aca5bf325" exitCode=0 Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.289141 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" event={"ID":"d725aabc-ba32-4c0e-bc91-8819d73cae40","Type":"ContainerDied","Data":"90f0ef44d50d3f68a6e80b10a816b133ae6b897ca8881964ca82e79aca5bf325"} Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.289180 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" event={"ID":"d725aabc-ba32-4c0e-bc91-8819d73cae40","Type":"ContainerDied","Data":"5dc8047188086fad5c8d0b1fabab1913e5036d9e22f49d173fa13d6dd588e3c0"} Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.289188 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-rtnbf" Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.289202 4775 scope.go:117] "RemoveContainer" containerID="90f0ef44d50d3f68a6e80b10a816b133ae6b897ca8881964ca82e79aca5bf325" Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.324336 4775 scope.go:117] "RemoveContainer" containerID="e55382d5f315dca58dbeb0320372028174ccdbb56163037ad99b712395ff7974" Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.339060 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-dns-swift-storage-0\") pod \"d725aabc-ba32-4c0e-bc91-8819d73cae40\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.339192 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dl7xb\" (UniqueName: \"kubernetes.io/projected/d725aabc-ba32-4c0e-bc91-8819d73cae40-kube-api-access-dl7xb\") pod \"d725aabc-ba32-4c0e-bc91-8819d73cae40\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.339260 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-ovsdbserver-nb\") pod \"d725aabc-ba32-4c0e-bc91-8819d73cae40\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.339337 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-dns-svc\") pod \"d725aabc-ba32-4c0e-bc91-8819d73cae40\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.339385 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-ovsdbserver-sb\") pod \"d725aabc-ba32-4c0e-bc91-8819d73cae40\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.339434 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-config\") pod \"d725aabc-ba32-4c0e-bc91-8819d73cae40\" (UID: \"d725aabc-ba32-4c0e-bc91-8819d73cae40\") " Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.366531 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d725aabc-ba32-4c0e-bc91-8819d73cae40-kube-api-access-dl7xb" (OuterVolumeSpecName: "kube-api-access-dl7xb") pod "d725aabc-ba32-4c0e-bc91-8819d73cae40" (UID: "d725aabc-ba32-4c0e-bc91-8819d73cae40"). InnerVolumeSpecName "kube-api-access-dl7xb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.398802 4775 scope.go:117] "RemoveContainer" containerID="90f0ef44d50d3f68a6e80b10a816b133ae6b897ca8881964ca82e79aca5bf325" Mar 21 05:11:06 crc kubenswrapper[4775]: E0321 05:11:06.399291 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90f0ef44d50d3f68a6e80b10a816b133ae6b897ca8881964ca82e79aca5bf325\": container with ID starting with 90f0ef44d50d3f68a6e80b10a816b133ae6b897ca8881964ca82e79aca5bf325 not found: ID does not exist" containerID="90f0ef44d50d3f68a6e80b10a816b133ae6b897ca8881964ca82e79aca5bf325" Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.399322 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90f0ef44d50d3f68a6e80b10a816b133ae6b897ca8881964ca82e79aca5bf325"} err="failed to get container status \"90f0ef44d50d3f68a6e80b10a816b133ae6b897ca8881964ca82e79aca5bf325\": rpc error: code = NotFound desc = could not find container \"90f0ef44d50d3f68a6e80b10a816b133ae6b897ca8881964ca82e79aca5bf325\": container with ID starting with 90f0ef44d50d3f68a6e80b10a816b133ae6b897ca8881964ca82e79aca5bf325 not found: ID does not exist" Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.399350 4775 scope.go:117] "RemoveContainer" containerID="e55382d5f315dca58dbeb0320372028174ccdbb56163037ad99b712395ff7974" Mar 21 05:11:06 crc kubenswrapper[4775]: E0321 05:11:06.400700 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e55382d5f315dca58dbeb0320372028174ccdbb56163037ad99b712395ff7974\": container with ID starting with e55382d5f315dca58dbeb0320372028174ccdbb56163037ad99b712395ff7974 not found: ID does not exist" containerID="e55382d5f315dca58dbeb0320372028174ccdbb56163037ad99b712395ff7974" Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.400776 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e55382d5f315dca58dbeb0320372028174ccdbb56163037ad99b712395ff7974"} err="failed to get container status \"e55382d5f315dca58dbeb0320372028174ccdbb56163037ad99b712395ff7974\": rpc error: code = NotFound desc = could not find container \"e55382d5f315dca58dbeb0320372028174ccdbb56163037ad99b712395ff7974\": container with ID starting with e55382d5f315dca58dbeb0320372028174ccdbb56163037ad99b712395ff7974 not found: ID does not exist" Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.426378 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d725aabc-ba32-4c0e-bc91-8819d73cae40" (UID: "d725aabc-ba32-4c0e-bc91-8819d73cae40"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.435379 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d725aabc-ba32-4c0e-bc91-8819d73cae40" (UID: "d725aabc-ba32-4c0e-bc91-8819d73cae40"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.441597 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-config" (OuterVolumeSpecName: "config") pod "d725aabc-ba32-4c0e-bc91-8819d73cae40" (UID: "d725aabc-ba32-4c0e-bc91-8819d73cae40"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.444327 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d725aabc-ba32-4c0e-bc91-8819d73cae40" (UID: "d725aabc-ba32-4c0e-bc91-8819d73cae40"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.445517 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.445538 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dl7xb\" (UniqueName: \"kubernetes.io/projected/d725aabc-ba32-4c0e-bc91-8819d73cae40-kube-api-access-dl7xb\") on node \"crc\" DevicePath \"\"" Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.445548 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.445558 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.445567 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.450789 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d725aabc-ba32-4c0e-bc91-8819d73cae40" (UID: "d725aabc-ba32-4c0e-bc91-8819d73cae40"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.546763 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d725aabc-ba32-4c0e-bc91-8819d73cae40-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.622520 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-rtnbf"] Mar 21 05:11:06 crc kubenswrapper[4775]: I0321 05:11:06.630530 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-rtnbf"] Mar 21 05:11:07 crc kubenswrapper[4775]: I0321 05:11:07.673475 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d725aabc-ba32-4c0e-bc91-8819d73cae40" path="/var/lib/kubelet/pods/d725aabc-ba32-4c0e-bc91-8819d73cae40/volumes" Mar 21 05:11:09 crc kubenswrapper[4775]: I0321 05:11:09.369685 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.125779 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m"] Mar 21 05:11:14 crc kubenswrapper[4775]: E0321 05:11:14.127530 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d725aabc-ba32-4c0e-bc91-8819d73cae40" containerName="dnsmasq-dns" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.127553 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d725aabc-ba32-4c0e-bc91-8819d73cae40" containerName="dnsmasq-dns" Mar 21 05:11:14 crc kubenswrapper[4775]: E0321 05:11:14.127624 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d725aabc-ba32-4c0e-bc91-8819d73cae40" containerName="init" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.127639 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d725aabc-ba32-4c0e-bc91-8819d73cae40" containerName="init" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.127928 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d725aabc-ba32-4c0e-bc91-8819d73cae40" containerName="dnsmasq-dns" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.129216 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.133252 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.133574 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.133765 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.133928 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljtt5" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.153167 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m"] Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.288400 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc7mq\" (UniqueName: \"kubernetes.io/projected/f41367b2-433d-48f7-af75-575be4b318fc-kube-api-access-sc7mq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m\" (UID: \"f41367b2-433d-48f7-af75-575be4b318fc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.288478 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f41367b2-433d-48f7-af75-575be4b318fc-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m\" (UID: \"f41367b2-433d-48f7-af75-575be4b318fc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.288520 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f41367b2-433d-48f7-af75-575be4b318fc-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m\" (UID: \"f41367b2-433d-48f7-af75-575be4b318fc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.288640 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f41367b2-433d-48f7-af75-575be4b318fc-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m\" (UID: \"f41367b2-433d-48f7-af75-575be4b318fc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.389842 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f41367b2-433d-48f7-af75-575be4b318fc-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m\" (UID: \"f41367b2-433d-48f7-af75-575be4b318fc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.389928 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc7mq\" (UniqueName: \"kubernetes.io/projected/f41367b2-433d-48f7-af75-575be4b318fc-kube-api-access-sc7mq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m\" (UID: \"f41367b2-433d-48f7-af75-575be4b318fc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.389965 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f41367b2-433d-48f7-af75-575be4b318fc-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m\" (UID: \"f41367b2-433d-48f7-af75-575be4b318fc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.389998 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f41367b2-433d-48f7-af75-575be4b318fc-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m\" (UID: \"f41367b2-433d-48f7-af75-575be4b318fc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.395932 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f41367b2-433d-48f7-af75-575be4b318fc-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m\" (UID: \"f41367b2-433d-48f7-af75-575be4b318fc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.396146 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f41367b2-433d-48f7-af75-575be4b318fc-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m\" (UID: \"f41367b2-433d-48f7-af75-575be4b318fc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.399976 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f41367b2-433d-48f7-af75-575be4b318fc-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m\" (UID: \"f41367b2-433d-48f7-af75-575be4b318fc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.414577 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc7mq\" (UniqueName: \"kubernetes.io/projected/f41367b2-433d-48f7-af75-575be4b318fc-kube-api-access-sc7mq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m\" (UID: \"f41367b2-433d-48f7-af75-575be4b318fc\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" Mar 21 05:11:14 crc kubenswrapper[4775]: I0321 05:11:14.483563 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" Mar 21 05:11:15 crc kubenswrapper[4775]: I0321 05:11:15.293959 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m"] Mar 21 05:11:15 crc kubenswrapper[4775]: W0321 05:11:15.298523 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf41367b2_433d_48f7_af75_575be4b318fc.slice/crio-fc1e18fbb56f39a46d6d2570a607e697306df431bf99ea7b40225a84d7c29234 WatchSource:0}: Error finding container fc1e18fbb56f39a46d6d2570a607e697306df431bf99ea7b40225a84d7c29234: Status 404 returned error can't find the container with id fc1e18fbb56f39a46d6d2570a607e697306df431bf99ea7b40225a84d7c29234 Mar 21 05:11:15 crc kubenswrapper[4775]: I0321 05:11:15.371470 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" event={"ID":"f41367b2-433d-48f7-af75-575be4b318fc","Type":"ContainerStarted","Data":"fc1e18fbb56f39a46d6d2570a607e697306df431bf99ea7b40225a84d7c29234"} Mar 21 05:11:25 crc kubenswrapper[4775]: I0321 05:11:25.476785 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" event={"ID":"f41367b2-433d-48f7-af75-575be4b318fc","Type":"ContainerStarted","Data":"fedd5444d5690bdee4579c2ab7d7bd0312f0ec5b6822006b2d668946f12b04e5"} Mar 21 05:11:25 crc kubenswrapper[4775]: I0321 05:11:25.498554 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" podStartSLOduration=2.328201561 podStartE2EDuration="11.498520349s" podCreationTimestamp="2026-03-21 05:11:14 +0000 UTC" firstStartedPulling="2026-03-21 05:11:15.300219168 +0000 UTC m=+1428.276682792" lastFinishedPulling="2026-03-21 05:11:24.470537956 +0000 UTC m=+1437.447001580" observedRunningTime="2026-03-21 05:11:25.49606445 +0000 UTC m=+1438.472528074" watchObservedRunningTime="2026-03-21 05:11:25.498520349 +0000 UTC m=+1438.474983983" Mar 21 05:11:29 crc kubenswrapper[4775]: I0321 05:11:29.514696 4775 generic.go:334] "Generic (PLEG): container finished" podID="e5e83941-a38d-4ee9-b967-1dac69c5a55b" containerID="92cf1e237e929c497cbf8f575434685971036de8960a393e86900a1e6642f0ed" exitCode=0 Mar 21 05:11:29 crc kubenswrapper[4775]: I0321 05:11:29.514775 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e5e83941-a38d-4ee9-b967-1dac69c5a55b","Type":"ContainerDied","Data":"92cf1e237e929c497cbf8f575434685971036de8960a393e86900a1e6642f0ed"} Mar 21 05:11:30 crc kubenswrapper[4775]: I0321 05:11:30.527698 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e5e83941-a38d-4ee9-b967-1dac69c5a55b","Type":"ContainerStarted","Data":"6bacd31d0f13fb2a0595ae6cad760ba6c31c94539fc57bb47ede52a58f85d70a"} Mar 21 05:11:30 crc kubenswrapper[4775]: I0321 05:11:30.528316 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 21 05:11:30 crc kubenswrapper[4775]: I0321 05:11:30.529886 4775 generic.go:334] "Generic (PLEG): container finished" podID="c95486b5-f2ad-4098-912d-6749b329824b" containerID="f565924e64e065b4ec412c0136fb28d6de7cffeff14b255d7fd47acd3851a689" exitCode=0 Mar 21 05:11:30 crc kubenswrapper[4775]: I0321 05:11:30.529923 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c95486b5-f2ad-4098-912d-6749b329824b","Type":"ContainerDied","Data":"f565924e64e065b4ec412c0136fb28d6de7cffeff14b255d7fd47acd3851a689"} Mar 21 05:11:30 crc kubenswrapper[4775]: I0321 05:11:30.558333 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.558311213 podStartE2EDuration="36.558311213s" podCreationTimestamp="2026-03-21 05:10:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:11:30.551281145 +0000 UTC m=+1443.527744769" watchObservedRunningTime="2026-03-21 05:11:30.558311213 +0000 UTC m=+1443.534774847" Mar 21 05:11:31 crc kubenswrapper[4775]: I0321 05:11:31.542523 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c95486b5-f2ad-4098-912d-6749b329824b","Type":"ContainerStarted","Data":"b46026fb47cd615b91a045db81faedfd96537575f80e33a593af4cbe28d57fe6"} Mar 21 05:11:31 crc kubenswrapper[4775]: I0321 05:11:31.543425 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:11:31 crc kubenswrapper[4775]: I0321 05:11:31.567210 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.567189829 podStartE2EDuration="36.567189829s" podCreationTimestamp="2026-03-21 05:10:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:11:31.560594974 +0000 UTC m=+1444.537058608" watchObservedRunningTime="2026-03-21 05:11:31.567189829 +0000 UTC m=+1444.543653463" Mar 21 05:11:35 crc kubenswrapper[4775]: I0321 05:11:35.406908 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z6mp2"] Mar 21 05:11:35 crc kubenswrapper[4775]: I0321 05:11:35.409766 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z6mp2" Mar 21 05:11:35 crc kubenswrapper[4775]: I0321 05:11:35.430928 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z6mp2"] Mar 21 05:11:35 crc kubenswrapper[4775]: I0321 05:11:35.458220 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bddbf00-7ac4-4180-b958-c377ef640334-catalog-content\") pod \"redhat-operators-z6mp2\" (UID: \"2bddbf00-7ac4-4180-b958-c377ef640334\") " pod="openshift-marketplace/redhat-operators-z6mp2" Mar 21 05:11:35 crc kubenswrapper[4775]: I0321 05:11:35.458423 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdsjf\" (UniqueName: \"kubernetes.io/projected/2bddbf00-7ac4-4180-b958-c377ef640334-kube-api-access-pdsjf\") pod \"redhat-operators-z6mp2\" (UID: \"2bddbf00-7ac4-4180-b958-c377ef640334\") " pod="openshift-marketplace/redhat-operators-z6mp2" Mar 21 05:11:35 crc kubenswrapper[4775]: I0321 05:11:35.458467 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bddbf00-7ac4-4180-b958-c377ef640334-utilities\") pod \"redhat-operators-z6mp2\" (UID: \"2bddbf00-7ac4-4180-b958-c377ef640334\") " pod="openshift-marketplace/redhat-operators-z6mp2" Mar 21 05:11:35 crc kubenswrapper[4775]: I0321 05:11:35.561547 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdsjf\" (UniqueName: \"kubernetes.io/projected/2bddbf00-7ac4-4180-b958-c377ef640334-kube-api-access-pdsjf\") pod \"redhat-operators-z6mp2\" (UID: \"2bddbf00-7ac4-4180-b958-c377ef640334\") " pod="openshift-marketplace/redhat-operators-z6mp2" Mar 21 05:11:35 crc kubenswrapper[4775]: I0321 05:11:35.561627 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bddbf00-7ac4-4180-b958-c377ef640334-utilities\") pod \"redhat-operators-z6mp2\" (UID: \"2bddbf00-7ac4-4180-b958-c377ef640334\") " pod="openshift-marketplace/redhat-operators-z6mp2" Mar 21 05:11:35 crc kubenswrapper[4775]: I0321 05:11:35.561681 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bddbf00-7ac4-4180-b958-c377ef640334-catalog-content\") pod \"redhat-operators-z6mp2\" (UID: \"2bddbf00-7ac4-4180-b958-c377ef640334\") " pod="openshift-marketplace/redhat-operators-z6mp2" Mar 21 05:11:35 crc kubenswrapper[4775]: I0321 05:11:35.562254 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bddbf00-7ac4-4180-b958-c377ef640334-catalog-content\") pod \"redhat-operators-z6mp2\" (UID: \"2bddbf00-7ac4-4180-b958-c377ef640334\") " pod="openshift-marketplace/redhat-operators-z6mp2" Mar 21 05:11:35 crc kubenswrapper[4775]: I0321 05:11:35.562886 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bddbf00-7ac4-4180-b958-c377ef640334-utilities\") pod \"redhat-operators-z6mp2\" (UID: \"2bddbf00-7ac4-4180-b958-c377ef640334\") " pod="openshift-marketplace/redhat-operators-z6mp2" Mar 21 05:11:35 crc kubenswrapper[4775]: I0321 05:11:35.606497 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdsjf\" (UniqueName: \"kubernetes.io/projected/2bddbf00-7ac4-4180-b958-c377ef640334-kube-api-access-pdsjf\") pod \"redhat-operators-z6mp2\" (UID: \"2bddbf00-7ac4-4180-b958-c377ef640334\") " pod="openshift-marketplace/redhat-operators-z6mp2" Mar 21 05:11:35 crc kubenswrapper[4775]: I0321 05:11:35.733672 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z6mp2" Mar 21 05:11:36 crc kubenswrapper[4775]: I0321 05:11:36.204584 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z6mp2"] Mar 21 05:11:36 crc kubenswrapper[4775]: I0321 05:11:36.600126 4775 generic.go:334] "Generic (PLEG): container finished" podID="2bddbf00-7ac4-4180-b958-c377ef640334" containerID="86bb0734a78e9ce15241affa8ae7b160b9d5b0cb227f27eb1579587505a0ae1e" exitCode=0 Mar 21 05:11:36 crc kubenswrapper[4775]: I0321 05:11:36.600200 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6mp2" event={"ID":"2bddbf00-7ac4-4180-b958-c377ef640334","Type":"ContainerDied","Data":"86bb0734a78e9ce15241affa8ae7b160b9d5b0cb227f27eb1579587505a0ae1e"} Mar 21 05:11:36 crc kubenswrapper[4775]: I0321 05:11:36.600230 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6mp2" event={"ID":"2bddbf00-7ac4-4180-b958-c377ef640334","Type":"ContainerStarted","Data":"2332fb4c24ce5b7478674331300ff4e968647562e22dc21f81edba2ae13017f2"} Mar 21 05:11:36 crc kubenswrapper[4775]: I0321 05:11:36.601932 4775 generic.go:334] "Generic (PLEG): container finished" podID="f41367b2-433d-48f7-af75-575be4b318fc" containerID="fedd5444d5690bdee4579c2ab7d7bd0312f0ec5b6822006b2d668946f12b04e5" exitCode=0 Mar 21 05:11:36 crc kubenswrapper[4775]: I0321 05:11:36.601975 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" event={"ID":"f41367b2-433d-48f7-af75-575be4b318fc","Type":"ContainerDied","Data":"fedd5444d5690bdee4579c2ab7d7bd0312f0ec5b6822006b2d668946f12b04e5"} Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.050407 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.209041 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f41367b2-433d-48f7-af75-575be4b318fc-ssh-key-openstack-edpm-ipam\") pod \"f41367b2-433d-48f7-af75-575be4b318fc\" (UID: \"f41367b2-433d-48f7-af75-575be4b318fc\") " Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.209158 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f41367b2-433d-48f7-af75-575be4b318fc-inventory\") pod \"f41367b2-433d-48f7-af75-575be4b318fc\" (UID: \"f41367b2-433d-48f7-af75-575be4b318fc\") " Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.209201 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f41367b2-433d-48f7-af75-575be4b318fc-repo-setup-combined-ca-bundle\") pod \"f41367b2-433d-48f7-af75-575be4b318fc\" (UID: \"f41367b2-433d-48f7-af75-575be4b318fc\") " Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.209280 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sc7mq\" (UniqueName: \"kubernetes.io/projected/f41367b2-433d-48f7-af75-575be4b318fc-kube-api-access-sc7mq\") pod \"f41367b2-433d-48f7-af75-575be4b318fc\" (UID: \"f41367b2-433d-48f7-af75-575be4b318fc\") " Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.226892 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f41367b2-433d-48f7-af75-575be4b318fc-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "f41367b2-433d-48f7-af75-575be4b318fc" (UID: "f41367b2-433d-48f7-af75-575be4b318fc"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.230060 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f41367b2-433d-48f7-af75-575be4b318fc-kube-api-access-sc7mq" (OuterVolumeSpecName: "kube-api-access-sc7mq") pod "f41367b2-433d-48f7-af75-575be4b318fc" (UID: "f41367b2-433d-48f7-af75-575be4b318fc"). InnerVolumeSpecName "kube-api-access-sc7mq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.242080 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f41367b2-433d-48f7-af75-575be4b318fc-inventory" (OuterVolumeSpecName: "inventory") pod "f41367b2-433d-48f7-af75-575be4b318fc" (UID: "f41367b2-433d-48f7-af75-575be4b318fc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.248656 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f41367b2-433d-48f7-af75-575be4b318fc-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f41367b2-433d-48f7-af75-575be4b318fc" (UID: "f41367b2-433d-48f7-af75-575be4b318fc"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.311302 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f41367b2-433d-48f7-af75-575be4b318fc-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.311350 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f41367b2-433d-48f7-af75-575be4b318fc-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.311364 4775 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f41367b2-433d-48f7-af75-575be4b318fc-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.311378 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sc7mq\" (UniqueName: \"kubernetes.io/projected/f41367b2-433d-48f7-af75-575be4b318fc-kube-api-access-sc7mq\") on node \"crc\" DevicePath \"\"" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.621196 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6mp2" event={"ID":"2bddbf00-7ac4-4180-b958-c377ef640334","Type":"ContainerStarted","Data":"04d5d232cc45dcabc0c4f738da19b014f8c8b51d249691a8d9e6bed3ed0986cd"} Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.624747 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" event={"ID":"f41367b2-433d-48f7-af75-575be4b318fc","Type":"ContainerDied","Data":"fc1e18fbb56f39a46d6d2570a607e697306df431bf99ea7b40225a84d7c29234"} Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.624784 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc1e18fbb56f39a46d6d2570a607e697306df431bf99ea7b40225a84d7c29234" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.624826 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.729517 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc"] Mar 21 05:11:38 crc kubenswrapper[4775]: E0321 05:11:38.729967 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f41367b2-433d-48f7-af75-575be4b318fc" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.729992 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f41367b2-433d-48f7-af75-575be4b318fc" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.730208 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f41367b2-433d-48f7-af75-575be4b318fc" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.730823 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.734794 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.734869 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.735010 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.735057 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljtt5" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.745602 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc"] Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.823465 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7c88e417-9ede-41d8-8337-79620ceb7798-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qgpzc\" (UID: \"7c88e417-9ede-41d8-8337-79620ceb7798\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.823682 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c88e417-9ede-41d8-8337-79620ceb7798-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qgpzc\" (UID: \"7c88e417-9ede-41d8-8337-79620ceb7798\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.823895 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dhq4\" (UniqueName: \"kubernetes.io/projected/7c88e417-9ede-41d8-8337-79620ceb7798-kube-api-access-7dhq4\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qgpzc\" (UID: \"7c88e417-9ede-41d8-8337-79620ceb7798\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.926166 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dhq4\" (UniqueName: \"kubernetes.io/projected/7c88e417-9ede-41d8-8337-79620ceb7798-kube-api-access-7dhq4\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qgpzc\" (UID: \"7c88e417-9ede-41d8-8337-79620ceb7798\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.926230 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7c88e417-9ede-41d8-8337-79620ceb7798-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qgpzc\" (UID: \"7c88e417-9ede-41d8-8337-79620ceb7798\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.927026 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c88e417-9ede-41d8-8337-79620ceb7798-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qgpzc\" (UID: \"7c88e417-9ede-41d8-8337-79620ceb7798\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.931206 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c88e417-9ede-41d8-8337-79620ceb7798-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qgpzc\" (UID: \"7c88e417-9ede-41d8-8337-79620ceb7798\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.931515 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7c88e417-9ede-41d8-8337-79620ceb7798-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qgpzc\" (UID: \"7c88e417-9ede-41d8-8337-79620ceb7798\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc" Mar 21 05:11:38 crc kubenswrapper[4775]: I0321 05:11:38.942629 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dhq4\" (UniqueName: \"kubernetes.io/projected/7c88e417-9ede-41d8-8337-79620ceb7798-kube-api-access-7dhq4\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qgpzc\" (UID: \"7c88e417-9ede-41d8-8337-79620ceb7798\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc" Mar 21 05:11:39 crc kubenswrapper[4775]: I0321 05:11:39.068607 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc" Mar 21 05:11:39 crc kubenswrapper[4775]: I0321 05:11:39.587255 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc"] Mar 21 05:11:39 crc kubenswrapper[4775]: I0321 05:11:39.637284 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc" event={"ID":"7c88e417-9ede-41d8-8337-79620ceb7798","Type":"ContainerStarted","Data":"78ac04b1c27fe6ac462f433d708306d43b03e18067a1c937699c9d1ee163b33a"} Mar 21 05:11:41 crc kubenswrapper[4775]: I0321 05:11:41.659232 4775 generic.go:334] "Generic (PLEG): container finished" podID="2bddbf00-7ac4-4180-b958-c377ef640334" containerID="04d5d232cc45dcabc0c4f738da19b014f8c8b51d249691a8d9e6bed3ed0986cd" exitCode=0 Mar 21 05:11:41 crc kubenswrapper[4775]: I0321 05:11:41.659278 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6mp2" event={"ID":"2bddbf00-7ac4-4180-b958-c377ef640334","Type":"ContainerDied","Data":"04d5d232cc45dcabc0c4f738da19b014f8c8b51d249691a8d9e6bed3ed0986cd"} Mar 21 05:11:42 crc kubenswrapper[4775]: I0321 05:11:42.787027 4775 scope.go:117] "RemoveContainer" containerID="9a3778eec58b097b77fc9e80249e34ac71c57ec91481179476b24785158c24e3" Mar 21 05:11:43 crc kubenswrapper[4775]: I0321 05:11:43.473161 4775 scope.go:117] "RemoveContainer" containerID="f749dd6bac424eb0695958e8b0d01fc15e3e824b72ea62322c8b3aee65b4d200" Mar 21 05:11:44 crc kubenswrapper[4775]: I0321 05:11:44.574715 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="e5e83941-a38d-4ee9-b967-1dac69c5a55b" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.211:5671: connect: connection refused" Mar 21 05:11:44 crc kubenswrapper[4775]: I0321 05:11:44.690890 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc" event={"ID":"7c88e417-9ede-41d8-8337-79620ceb7798","Type":"ContainerStarted","Data":"e57bfbd0e9b0250d0a5dde317af9eac11f2c7f53c320d67ebd8107bb4429316b"} Mar 21 05:11:44 crc kubenswrapper[4775]: I0321 05:11:44.694910 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6mp2" event={"ID":"2bddbf00-7ac4-4180-b958-c377ef640334","Type":"ContainerStarted","Data":"c9ab0fd1a0cc329833eab925ace21a193f7a42c34989b229a72bfaf73bdc5341"} Mar 21 05:11:44 crc kubenswrapper[4775]: I0321 05:11:44.717356 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc" podStartSLOduration=2.476692311 podStartE2EDuration="6.717331494s" podCreationTimestamp="2026-03-21 05:11:38 +0000 UTC" firstStartedPulling="2026-03-21 05:11:39.596070292 +0000 UTC m=+1452.572533936" lastFinishedPulling="2026-03-21 05:11:43.836709485 +0000 UTC m=+1456.813173119" observedRunningTime="2026-03-21 05:11:44.711556362 +0000 UTC m=+1457.688020006" watchObservedRunningTime="2026-03-21 05:11:44.717331494 +0000 UTC m=+1457.693795118" Mar 21 05:11:44 crc kubenswrapper[4775]: I0321 05:11:44.741834 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z6mp2" podStartSLOduration=2.087083223 podStartE2EDuration="9.741813703s" podCreationTimestamp="2026-03-21 05:11:35 +0000 UTC" firstStartedPulling="2026-03-21 05:11:36.601536196 +0000 UTC m=+1449.577999820" lastFinishedPulling="2026-03-21 05:11:44.256266676 +0000 UTC m=+1457.232730300" observedRunningTime="2026-03-21 05:11:44.739473847 +0000 UTC m=+1457.715937471" watchObservedRunningTime="2026-03-21 05:11:44.741813703 +0000 UTC m=+1457.718277347" Mar 21 05:11:45 crc kubenswrapper[4775]: I0321 05:11:45.672170 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 21 05:11:45 crc kubenswrapper[4775]: I0321 05:11:45.734495 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z6mp2" Mar 21 05:11:45 crc kubenswrapper[4775]: I0321 05:11:45.734744 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z6mp2" Mar 21 05:11:46 crc kubenswrapper[4775]: I0321 05:11:46.715707 4775 generic.go:334] "Generic (PLEG): container finished" podID="7c88e417-9ede-41d8-8337-79620ceb7798" containerID="e57bfbd0e9b0250d0a5dde317af9eac11f2c7f53c320d67ebd8107bb4429316b" exitCode=0 Mar 21 05:11:46 crc kubenswrapper[4775]: I0321 05:11:46.715798 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc" event={"ID":"7c88e417-9ede-41d8-8337-79620ceb7798","Type":"ContainerDied","Data":"e57bfbd0e9b0250d0a5dde317af9eac11f2c7f53c320d67ebd8107bb4429316b"} Mar 21 05:11:46 crc kubenswrapper[4775]: I0321 05:11:46.804786 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-z6mp2" podUID="2bddbf00-7ac4-4180-b958-c377ef640334" containerName="registry-server" probeResult="failure" output=< Mar 21 05:11:46 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Mar 21 05:11:46 crc kubenswrapper[4775]: > Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.179238 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc" Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.308566 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c88e417-9ede-41d8-8337-79620ceb7798-inventory\") pod \"7c88e417-9ede-41d8-8337-79620ceb7798\" (UID: \"7c88e417-9ede-41d8-8337-79620ceb7798\") " Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.308621 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dhq4\" (UniqueName: \"kubernetes.io/projected/7c88e417-9ede-41d8-8337-79620ceb7798-kube-api-access-7dhq4\") pod \"7c88e417-9ede-41d8-8337-79620ceb7798\" (UID: \"7c88e417-9ede-41d8-8337-79620ceb7798\") " Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.308766 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7c88e417-9ede-41d8-8337-79620ceb7798-ssh-key-openstack-edpm-ipam\") pod \"7c88e417-9ede-41d8-8337-79620ceb7798\" (UID: \"7c88e417-9ede-41d8-8337-79620ceb7798\") " Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.315582 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c88e417-9ede-41d8-8337-79620ceb7798-kube-api-access-7dhq4" (OuterVolumeSpecName: "kube-api-access-7dhq4") pod "7c88e417-9ede-41d8-8337-79620ceb7798" (UID: "7c88e417-9ede-41d8-8337-79620ceb7798"). InnerVolumeSpecName "kube-api-access-7dhq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.342287 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c88e417-9ede-41d8-8337-79620ceb7798-inventory" (OuterVolumeSpecName: "inventory") pod "7c88e417-9ede-41d8-8337-79620ceb7798" (UID: "7c88e417-9ede-41d8-8337-79620ceb7798"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.346828 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c88e417-9ede-41d8-8337-79620ceb7798-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7c88e417-9ede-41d8-8337-79620ceb7798" (UID: "7c88e417-9ede-41d8-8337-79620ceb7798"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.410681 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7c88e417-9ede-41d8-8337-79620ceb7798-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.410714 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c88e417-9ede-41d8-8337-79620ceb7798-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.410726 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dhq4\" (UniqueName: \"kubernetes.io/projected/7c88e417-9ede-41d8-8337-79620ceb7798-kube-api-access-7dhq4\") on node \"crc\" DevicePath \"\"" Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.736529 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc" event={"ID":"7c88e417-9ede-41d8-8337-79620ceb7798","Type":"ContainerDied","Data":"78ac04b1c27fe6ac462f433d708306d43b03e18067a1c937699c9d1ee163b33a"} Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.736957 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78ac04b1c27fe6ac462f433d708306d43b03e18067a1c937699c9d1ee163b33a" Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.736788 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgpzc" Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.834957 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m"] Mar 21 05:11:48 crc kubenswrapper[4775]: E0321 05:11:48.835345 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c88e417-9ede-41d8-8337-79620ceb7798" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.835362 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c88e417-9ede-41d8-8337-79620ceb7798" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.835546 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c88e417-9ede-41d8-8337-79620ceb7798" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.836203 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.840515 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.844829 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.845364 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.848567 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m"] Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.854963 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljtt5" Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.923815 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/203df932-0574-4098-b897-ba50813f2ec1-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m\" (UID: \"203df932-0574-4098-b897-ba50813f2ec1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.924168 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/203df932-0574-4098-b897-ba50813f2ec1-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m\" (UID: \"203df932-0574-4098-b897-ba50813f2ec1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.924290 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/203df932-0574-4098-b897-ba50813f2ec1-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m\" (UID: \"203df932-0574-4098-b897-ba50813f2ec1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" Mar 21 05:11:48 crc kubenswrapper[4775]: I0321 05:11:48.924361 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sfxh\" (UniqueName: \"kubernetes.io/projected/203df932-0574-4098-b897-ba50813f2ec1-kube-api-access-5sfxh\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m\" (UID: \"203df932-0574-4098-b897-ba50813f2ec1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" Mar 21 05:11:49 crc kubenswrapper[4775]: I0321 05:11:49.025958 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/203df932-0574-4098-b897-ba50813f2ec1-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m\" (UID: \"203df932-0574-4098-b897-ba50813f2ec1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" Mar 21 05:11:49 crc kubenswrapper[4775]: I0321 05:11:49.027153 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/203df932-0574-4098-b897-ba50813f2ec1-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m\" (UID: \"203df932-0574-4098-b897-ba50813f2ec1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" Mar 21 05:11:49 crc kubenswrapper[4775]: I0321 05:11:49.027305 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/203df932-0574-4098-b897-ba50813f2ec1-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m\" (UID: \"203df932-0574-4098-b897-ba50813f2ec1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" Mar 21 05:11:49 crc kubenswrapper[4775]: I0321 05:11:49.027401 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sfxh\" (UniqueName: \"kubernetes.io/projected/203df932-0574-4098-b897-ba50813f2ec1-kube-api-access-5sfxh\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m\" (UID: \"203df932-0574-4098-b897-ba50813f2ec1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" Mar 21 05:11:49 crc kubenswrapper[4775]: I0321 05:11:49.031198 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/203df932-0574-4098-b897-ba50813f2ec1-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m\" (UID: \"203df932-0574-4098-b897-ba50813f2ec1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" Mar 21 05:11:49 crc kubenswrapper[4775]: I0321 05:11:49.031321 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/203df932-0574-4098-b897-ba50813f2ec1-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m\" (UID: \"203df932-0574-4098-b897-ba50813f2ec1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" Mar 21 05:11:49 crc kubenswrapper[4775]: I0321 05:11:49.038552 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/203df932-0574-4098-b897-ba50813f2ec1-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m\" (UID: \"203df932-0574-4098-b897-ba50813f2ec1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" Mar 21 05:11:49 crc kubenswrapper[4775]: I0321 05:11:49.075813 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sfxh\" (UniqueName: \"kubernetes.io/projected/203df932-0574-4098-b897-ba50813f2ec1-kube-api-access-5sfxh\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m\" (UID: \"203df932-0574-4098-b897-ba50813f2ec1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" Mar 21 05:11:49 crc kubenswrapper[4775]: I0321 05:11:49.156217 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" Mar 21 05:11:49 crc kubenswrapper[4775]: I0321 05:11:49.701361 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m"] Mar 21 05:11:49 crc kubenswrapper[4775]: I0321 05:11:49.750876 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" event={"ID":"203df932-0574-4098-b897-ba50813f2ec1","Type":"ContainerStarted","Data":"3c801bc6a121fd4faebee72f398b0bcdeabcebcace5ab4c44442f937b77950dc"} Mar 21 05:11:50 crc kubenswrapper[4775]: I0321 05:11:50.760485 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" event={"ID":"203df932-0574-4098-b897-ba50813f2ec1","Type":"ContainerStarted","Data":"e83db7ccdc230f6a7c372f2ab3e28ef862efd0646a039a1d8f53714e19540048"} Mar 21 05:11:50 crc kubenswrapper[4775]: I0321 05:11:50.784521 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" podStartSLOduration=2.330224394 podStartE2EDuration="2.784497711s" podCreationTimestamp="2026-03-21 05:11:48 +0000 UTC" firstStartedPulling="2026-03-21 05:11:49.703557788 +0000 UTC m=+1462.680021412" lastFinishedPulling="2026-03-21 05:11:50.157831105 +0000 UTC m=+1463.134294729" observedRunningTime="2026-03-21 05:11:50.780098807 +0000 UTC m=+1463.756562431" watchObservedRunningTime="2026-03-21 05:11:50.784497711 +0000 UTC m=+1463.760961335" Mar 21 05:11:54 crc kubenswrapper[4775]: I0321 05:11:54.577946 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 21 05:11:55 crc kubenswrapper[4775]: I0321 05:11:55.796019 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z6mp2" Mar 21 05:11:55 crc kubenswrapper[4775]: I0321 05:11:55.850330 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z6mp2" Mar 21 05:11:56 crc kubenswrapper[4775]: I0321 05:11:56.029149 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z6mp2"] Mar 21 05:11:56 crc kubenswrapper[4775]: I0321 05:11:56.836268 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z6mp2" podUID="2bddbf00-7ac4-4180-b958-c377ef640334" containerName="registry-server" containerID="cri-o://c9ab0fd1a0cc329833eab925ace21a193f7a42c34989b229a72bfaf73bdc5341" gracePeriod=2 Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.250075 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z6mp2" Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.401667 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bddbf00-7ac4-4180-b958-c377ef640334-catalog-content\") pod \"2bddbf00-7ac4-4180-b958-c377ef640334\" (UID: \"2bddbf00-7ac4-4180-b958-c377ef640334\") " Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.401804 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bddbf00-7ac4-4180-b958-c377ef640334-utilities\") pod \"2bddbf00-7ac4-4180-b958-c377ef640334\" (UID: \"2bddbf00-7ac4-4180-b958-c377ef640334\") " Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.402005 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdsjf\" (UniqueName: \"kubernetes.io/projected/2bddbf00-7ac4-4180-b958-c377ef640334-kube-api-access-pdsjf\") pod \"2bddbf00-7ac4-4180-b958-c377ef640334\" (UID: \"2bddbf00-7ac4-4180-b958-c377ef640334\") " Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.402754 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bddbf00-7ac4-4180-b958-c377ef640334-utilities" (OuterVolumeSpecName: "utilities") pod "2bddbf00-7ac4-4180-b958-c377ef640334" (UID: "2bddbf00-7ac4-4180-b958-c377ef640334"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.408203 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bddbf00-7ac4-4180-b958-c377ef640334-kube-api-access-pdsjf" (OuterVolumeSpecName: "kube-api-access-pdsjf") pod "2bddbf00-7ac4-4180-b958-c377ef640334" (UID: "2bddbf00-7ac4-4180-b958-c377ef640334"). InnerVolumeSpecName "kube-api-access-pdsjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.504278 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdsjf\" (UniqueName: \"kubernetes.io/projected/2bddbf00-7ac4-4180-b958-c377ef640334-kube-api-access-pdsjf\") on node \"crc\" DevicePath \"\"" Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.504319 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bddbf00-7ac4-4180-b958-c377ef640334-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.525417 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bddbf00-7ac4-4180-b958-c377ef640334-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2bddbf00-7ac4-4180-b958-c377ef640334" (UID: "2bddbf00-7ac4-4180-b958-c377ef640334"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.606790 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bddbf00-7ac4-4180-b958-c377ef640334-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.860006 4775 generic.go:334] "Generic (PLEG): container finished" podID="2bddbf00-7ac4-4180-b958-c377ef640334" containerID="c9ab0fd1a0cc329833eab925ace21a193f7a42c34989b229a72bfaf73bdc5341" exitCode=0 Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.860463 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6mp2" event={"ID":"2bddbf00-7ac4-4180-b958-c377ef640334","Type":"ContainerDied","Data":"c9ab0fd1a0cc329833eab925ace21a193f7a42c34989b229a72bfaf73bdc5341"} Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.860509 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6mp2" event={"ID":"2bddbf00-7ac4-4180-b958-c377ef640334","Type":"ContainerDied","Data":"2332fb4c24ce5b7478674331300ff4e968647562e22dc21f81edba2ae13017f2"} Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.860555 4775 scope.go:117] "RemoveContainer" containerID="c9ab0fd1a0cc329833eab925ace21a193f7a42c34989b229a72bfaf73bdc5341" Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.860879 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z6mp2" Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.903768 4775 scope.go:117] "RemoveContainer" containerID="04d5d232cc45dcabc0c4f738da19b014f8c8b51d249691a8d9e6bed3ed0986cd" Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.903922 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z6mp2"] Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.911020 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z6mp2"] Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.937099 4775 scope.go:117] "RemoveContainer" containerID="86bb0734a78e9ce15241affa8ae7b160b9d5b0cb227f27eb1579587505a0ae1e" Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.975872 4775 scope.go:117] "RemoveContainer" containerID="c9ab0fd1a0cc329833eab925ace21a193f7a42c34989b229a72bfaf73bdc5341" Mar 21 05:11:57 crc kubenswrapper[4775]: E0321 05:11:57.976230 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9ab0fd1a0cc329833eab925ace21a193f7a42c34989b229a72bfaf73bdc5341\": container with ID starting with c9ab0fd1a0cc329833eab925ace21a193f7a42c34989b229a72bfaf73bdc5341 not found: ID does not exist" containerID="c9ab0fd1a0cc329833eab925ace21a193f7a42c34989b229a72bfaf73bdc5341" Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.976267 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9ab0fd1a0cc329833eab925ace21a193f7a42c34989b229a72bfaf73bdc5341"} err="failed to get container status \"c9ab0fd1a0cc329833eab925ace21a193f7a42c34989b229a72bfaf73bdc5341\": rpc error: code = NotFound desc = could not find container \"c9ab0fd1a0cc329833eab925ace21a193f7a42c34989b229a72bfaf73bdc5341\": container with ID starting with c9ab0fd1a0cc329833eab925ace21a193f7a42c34989b229a72bfaf73bdc5341 not found: ID does not exist" Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.976288 4775 scope.go:117] "RemoveContainer" containerID="04d5d232cc45dcabc0c4f738da19b014f8c8b51d249691a8d9e6bed3ed0986cd" Mar 21 05:11:57 crc kubenswrapper[4775]: E0321 05:11:57.976555 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04d5d232cc45dcabc0c4f738da19b014f8c8b51d249691a8d9e6bed3ed0986cd\": container with ID starting with 04d5d232cc45dcabc0c4f738da19b014f8c8b51d249691a8d9e6bed3ed0986cd not found: ID does not exist" containerID="04d5d232cc45dcabc0c4f738da19b014f8c8b51d249691a8d9e6bed3ed0986cd" Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.976581 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04d5d232cc45dcabc0c4f738da19b014f8c8b51d249691a8d9e6bed3ed0986cd"} err="failed to get container status \"04d5d232cc45dcabc0c4f738da19b014f8c8b51d249691a8d9e6bed3ed0986cd\": rpc error: code = NotFound desc = could not find container \"04d5d232cc45dcabc0c4f738da19b014f8c8b51d249691a8d9e6bed3ed0986cd\": container with ID starting with 04d5d232cc45dcabc0c4f738da19b014f8c8b51d249691a8d9e6bed3ed0986cd not found: ID does not exist" Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.976595 4775 scope.go:117] "RemoveContainer" containerID="86bb0734a78e9ce15241affa8ae7b160b9d5b0cb227f27eb1579587505a0ae1e" Mar 21 05:11:57 crc kubenswrapper[4775]: E0321 05:11:57.976831 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86bb0734a78e9ce15241affa8ae7b160b9d5b0cb227f27eb1579587505a0ae1e\": container with ID starting with 86bb0734a78e9ce15241affa8ae7b160b9d5b0cb227f27eb1579587505a0ae1e not found: ID does not exist" containerID="86bb0734a78e9ce15241affa8ae7b160b9d5b0cb227f27eb1579587505a0ae1e" Mar 21 05:11:57 crc kubenswrapper[4775]: I0321 05:11:57.976885 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86bb0734a78e9ce15241affa8ae7b160b9d5b0cb227f27eb1579587505a0ae1e"} err="failed to get container status \"86bb0734a78e9ce15241affa8ae7b160b9d5b0cb227f27eb1579587505a0ae1e\": rpc error: code = NotFound desc = could not find container \"86bb0734a78e9ce15241affa8ae7b160b9d5b0cb227f27eb1579587505a0ae1e\": container with ID starting with 86bb0734a78e9ce15241affa8ae7b160b9d5b0cb227f27eb1579587505a0ae1e not found: ID does not exist" Mar 21 05:11:59 crc kubenswrapper[4775]: I0321 05:11:59.672236 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bddbf00-7ac4-4180-b958-c377ef640334" path="/var/lib/kubelet/pods/2bddbf00-7ac4-4180-b958-c377ef640334/volumes" Mar 21 05:12:00 crc kubenswrapper[4775]: I0321 05:12:00.146851 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567832-cm8v6"] Mar 21 05:12:00 crc kubenswrapper[4775]: E0321 05:12:00.147510 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bddbf00-7ac4-4180-b958-c377ef640334" containerName="registry-server" Mar 21 05:12:00 crc kubenswrapper[4775]: I0321 05:12:00.147544 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bddbf00-7ac4-4180-b958-c377ef640334" containerName="registry-server" Mar 21 05:12:00 crc kubenswrapper[4775]: E0321 05:12:00.147603 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bddbf00-7ac4-4180-b958-c377ef640334" containerName="extract-content" Mar 21 05:12:00 crc kubenswrapper[4775]: I0321 05:12:00.147617 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bddbf00-7ac4-4180-b958-c377ef640334" containerName="extract-content" Mar 21 05:12:00 crc kubenswrapper[4775]: E0321 05:12:00.147644 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bddbf00-7ac4-4180-b958-c377ef640334" containerName="extract-utilities" Mar 21 05:12:00 crc kubenswrapper[4775]: I0321 05:12:00.147659 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bddbf00-7ac4-4180-b958-c377ef640334" containerName="extract-utilities" Mar 21 05:12:00 crc kubenswrapper[4775]: I0321 05:12:00.147980 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bddbf00-7ac4-4180-b958-c377ef640334" containerName="registry-server" Mar 21 05:12:00 crc kubenswrapper[4775]: I0321 05:12:00.149044 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567832-cm8v6" Mar 21 05:12:00 crc kubenswrapper[4775]: I0321 05:12:00.151037 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:12:00 crc kubenswrapper[4775]: I0321 05:12:00.151561 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:12:00 crc kubenswrapper[4775]: I0321 05:12:00.151788 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:12:00 crc kubenswrapper[4775]: I0321 05:12:00.172434 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567832-cm8v6"] Mar 21 05:12:00 crc kubenswrapper[4775]: I0321 05:12:00.257699 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrhr8\" (UniqueName: \"kubernetes.io/projected/0ffbca22-6533-4ade-8f1f-eeeda82f159f-kube-api-access-lrhr8\") pod \"auto-csr-approver-29567832-cm8v6\" (UID: \"0ffbca22-6533-4ade-8f1f-eeeda82f159f\") " pod="openshift-infra/auto-csr-approver-29567832-cm8v6" Mar 21 05:12:00 crc kubenswrapper[4775]: I0321 05:12:00.362241 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrhr8\" (UniqueName: \"kubernetes.io/projected/0ffbca22-6533-4ade-8f1f-eeeda82f159f-kube-api-access-lrhr8\") pod \"auto-csr-approver-29567832-cm8v6\" (UID: \"0ffbca22-6533-4ade-8f1f-eeeda82f159f\") " pod="openshift-infra/auto-csr-approver-29567832-cm8v6" Mar 21 05:12:00 crc kubenswrapper[4775]: I0321 05:12:00.387144 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrhr8\" (UniqueName: \"kubernetes.io/projected/0ffbca22-6533-4ade-8f1f-eeeda82f159f-kube-api-access-lrhr8\") pod \"auto-csr-approver-29567832-cm8v6\" (UID: \"0ffbca22-6533-4ade-8f1f-eeeda82f159f\") " pod="openshift-infra/auto-csr-approver-29567832-cm8v6" Mar 21 05:12:00 crc kubenswrapper[4775]: I0321 05:12:00.472173 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567832-cm8v6" Mar 21 05:12:01 crc kubenswrapper[4775]: I0321 05:12:01.163202 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567832-cm8v6"] Mar 21 05:12:01 crc kubenswrapper[4775]: I0321 05:12:01.904758 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567832-cm8v6" event={"ID":"0ffbca22-6533-4ade-8f1f-eeeda82f159f","Type":"ContainerStarted","Data":"613e7e7dde3564958173ee27dab8808a83afb18788c0f6e13f0edfacec7337ef"} Mar 21 05:12:02 crc kubenswrapper[4775]: I0321 05:12:02.914421 4775 generic.go:334] "Generic (PLEG): container finished" podID="0ffbca22-6533-4ade-8f1f-eeeda82f159f" containerID="fdab6584b0a9bddc76fc370c865954a1758e24e006007d8b2953684194b46540" exitCode=0 Mar 21 05:12:02 crc kubenswrapper[4775]: I0321 05:12:02.914460 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567832-cm8v6" event={"ID":"0ffbca22-6533-4ade-8f1f-eeeda82f159f","Type":"ContainerDied","Data":"fdab6584b0a9bddc76fc370c865954a1758e24e006007d8b2953684194b46540"} Mar 21 05:12:04 crc kubenswrapper[4775]: I0321 05:12:04.319930 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567832-cm8v6" Mar 21 05:12:04 crc kubenswrapper[4775]: I0321 05:12:04.388013 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrhr8\" (UniqueName: \"kubernetes.io/projected/0ffbca22-6533-4ade-8f1f-eeeda82f159f-kube-api-access-lrhr8\") pod \"0ffbca22-6533-4ade-8f1f-eeeda82f159f\" (UID: \"0ffbca22-6533-4ade-8f1f-eeeda82f159f\") " Mar 21 05:12:04 crc kubenswrapper[4775]: I0321 05:12:04.396756 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ffbca22-6533-4ade-8f1f-eeeda82f159f-kube-api-access-lrhr8" (OuterVolumeSpecName: "kube-api-access-lrhr8") pod "0ffbca22-6533-4ade-8f1f-eeeda82f159f" (UID: "0ffbca22-6533-4ade-8f1f-eeeda82f159f"). InnerVolumeSpecName "kube-api-access-lrhr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:12:04 crc kubenswrapper[4775]: I0321 05:12:04.490264 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrhr8\" (UniqueName: \"kubernetes.io/projected/0ffbca22-6533-4ade-8f1f-eeeda82f159f-kube-api-access-lrhr8\") on node \"crc\" DevicePath \"\"" Mar 21 05:12:04 crc kubenswrapper[4775]: I0321 05:12:04.933907 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567832-cm8v6" event={"ID":"0ffbca22-6533-4ade-8f1f-eeeda82f159f","Type":"ContainerDied","Data":"613e7e7dde3564958173ee27dab8808a83afb18788c0f6e13f0edfacec7337ef"} Mar 21 05:12:04 crc kubenswrapper[4775]: I0321 05:12:04.933952 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="613e7e7dde3564958173ee27dab8808a83afb18788c0f6e13f0edfacec7337ef" Mar 21 05:12:04 crc kubenswrapper[4775]: I0321 05:12:04.933955 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567832-cm8v6" Mar 21 05:12:05 crc kubenswrapper[4775]: I0321 05:12:05.390063 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567826-89d6v"] Mar 21 05:12:05 crc kubenswrapper[4775]: I0321 05:12:05.400863 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567826-89d6v"] Mar 21 05:12:05 crc kubenswrapper[4775]: I0321 05:12:05.686502 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c64d94e-917c-49b6-824b-0b2bdf9691ef" path="/var/lib/kubelet/pods/0c64d94e-917c-49b6-824b-0b2bdf9691ef/volumes" Mar 21 05:12:43 crc kubenswrapper[4775]: I0321 05:12:43.974976 4775 scope.go:117] "RemoveContainer" containerID="af6715536a4cd1315ee5f22687cb60f190dfcee964ee03a3fb9b9b38cd6a20b3" Mar 21 05:12:44 crc kubenswrapper[4775]: I0321 05:12:44.007495 4775 scope.go:117] "RemoveContainer" containerID="8177bcef8f42975589ec3469527727473adafaa3cc6210ffa908111b0912b0d5" Mar 21 05:12:44 crc kubenswrapper[4775]: I0321 05:12:44.049430 4775 scope.go:117] "RemoveContainer" containerID="2acf177da0f2ed7ac0606a427af701da980e32ea57fc8b3636ddd64f2a8a5536" Mar 21 05:13:02 crc kubenswrapper[4775]: I0321 05:13:02.482324 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:13:02 crc kubenswrapper[4775]: I0321 05:13:02.482914 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:13:15 crc kubenswrapper[4775]: I0321 05:13:15.962740 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6kpp4"] Mar 21 05:13:15 crc kubenswrapper[4775]: E0321 05:13:15.964043 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ffbca22-6533-4ade-8f1f-eeeda82f159f" containerName="oc" Mar 21 05:13:15 crc kubenswrapper[4775]: I0321 05:13:15.964066 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ffbca22-6533-4ade-8f1f-eeeda82f159f" containerName="oc" Mar 21 05:13:15 crc kubenswrapper[4775]: I0321 05:13:15.964416 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ffbca22-6533-4ade-8f1f-eeeda82f159f" containerName="oc" Mar 21 05:13:15 crc kubenswrapper[4775]: I0321 05:13:15.966973 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6kpp4" Mar 21 05:13:15 crc kubenswrapper[4775]: I0321 05:13:15.974454 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6kpp4"] Mar 21 05:13:16 crc kubenswrapper[4775]: I0321 05:13:16.143513 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45493002-fd0c-4afd-b556-846d8cac5002-catalog-content\") pod \"certified-operators-6kpp4\" (UID: \"45493002-fd0c-4afd-b556-846d8cac5002\") " pod="openshift-marketplace/certified-operators-6kpp4" Mar 21 05:13:16 crc kubenswrapper[4775]: I0321 05:13:16.143588 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rglxd\" (UniqueName: \"kubernetes.io/projected/45493002-fd0c-4afd-b556-846d8cac5002-kube-api-access-rglxd\") pod \"certified-operators-6kpp4\" (UID: \"45493002-fd0c-4afd-b556-846d8cac5002\") " pod="openshift-marketplace/certified-operators-6kpp4" Mar 21 05:13:16 crc kubenswrapper[4775]: I0321 05:13:16.143634 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45493002-fd0c-4afd-b556-846d8cac5002-utilities\") pod \"certified-operators-6kpp4\" (UID: \"45493002-fd0c-4afd-b556-846d8cac5002\") " pod="openshift-marketplace/certified-operators-6kpp4" Mar 21 05:13:16 crc kubenswrapper[4775]: I0321 05:13:16.245685 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45493002-fd0c-4afd-b556-846d8cac5002-catalog-content\") pod \"certified-operators-6kpp4\" (UID: \"45493002-fd0c-4afd-b556-846d8cac5002\") " pod="openshift-marketplace/certified-operators-6kpp4" Mar 21 05:13:16 crc kubenswrapper[4775]: I0321 05:13:16.245752 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rglxd\" (UniqueName: \"kubernetes.io/projected/45493002-fd0c-4afd-b556-846d8cac5002-kube-api-access-rglxd\") pod \"certified-operators-6kpp4\" (UID: \"45493002-fd0c-4afd-b556-846d8cac5002\") " pod="openshift-marketplace/certified-operators-6kpp4" Mar 21 05:13:16 crc kubenswrapper[4775]: I0321 05:13:16.245789 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45493002-fd0c-4afd-b556-846d8cac5002-utilities\") pod \"certified-operators-6kpp4\" (UID: \"45493002-fd0c-4afd-b556-846d8cac5002\") " pod="openshift-marketplace/certified-operators-6kpp4" Mar 21 05:13:16 crc kubenswrapper[4775]: I0321 05:13:16.246409 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45493002-fd0c-4afd-b556-846d8cac5002-catalog-content\") pod \"certified-operators-6kpp4\" (UID: \"45493002-fd0c-4afd-b556-846d8cac5002\") " pod="openshift-marketplace/certified-operators-6kpp4" Mar 21 05:13:16 crc kubenswrapper[4775]: I0321 05:13:16.246492 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45493002-fd0c-4afd-b556-846d8cac5002-utilities\") pod \"certified-operators-6kpp4\" (UID: \"45493002-fd0c-4afd-b556-846d8cac5002\") " pod="openshift-marketplace/certified-operators-6kpp4" Mar 21 05:13:16 crc kubenswrapper[4775]: I0321 05:13:16.271323 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rglxd\" (UniqueName: \"kubernetes.io/projected/45493002-fd0c-4afd-b556-846d8cac5002-kube-api-access-rglxd\") pod \"certified-operators-6kpp4\" (UID: \"45493002-fd0c-4afd-b556-846d8cac5002\") " pod="openshift-marketplace/certified-operators-6kpp4" Mar 21 05:13:16 crc kubenswrapper[4775]: I0321 05:13:16.286765 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6kpp4" Mar 21 05:13:16 crc kubenswrapper[4775]: I0321 05:13:16.918906 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6kpp4"] Mar 21 05:13:17 crc kubenswrapper[4775]: I0321 05:13:17.590686 4775 generic.go:334] "Generic (PLEG): container finished" podID="45493002-fd0c-4afd-b556-846d8cac5002" containerID="cdca173b9c8744ded95289b25e21f20b4cd227c1dd0864ebe8f6347181032830" exitCode=0 Mar 21 05:13:17 crc kubenswrapper[4775]: I0321 05:13:17.590987 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6kpp4" event={"ID":"45493002-fd0c-4afd-b556-846d8cac5002","Type":"ContainerDied","Data":"cdca173b9c8744ded95289b25e21f20b4cd227c1dd0864ebe8f6347181032830"} Mar 21 05:13:17 crc kubenswrapper[4775]: I0321 05:13:17.591013 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6kpp4" event={"ID":"45493002-fd0c-4afd-b556-846d8cac5002","Type":"ContainerStarted","Data":"f4eddf91f9b502f94f8ea62789271ca7b116cfd53376cf9077858bb64d6d2b97"} Mar 21 05:13:19 crc kubenswrapper[4775]: I0321 05:13:19.615352 4775 generic.go:334] "Generic (PLEG): container finished" podID="45493002-fd0c-4afd-b556-846d8cac5002" containerID="a6782ced417e87e9facb555ea87dc070f4238ca0a19d2f8a047d54825d30f5c4" exitCode=0 Mar 21 05:13:19 crc kubenswrapper[4775]: I0321 05:13:19.615416 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6kpp4" event={"ID":"45493002-fd0c-4afd-b556-846d8cac5002","Type":"ContainerDied","Data":"a6782ced417e87e9facb555ea87dc070f4238ca0a19d2f8a047d54825d30f5c4"} Mar 21 05:13:21 crc kubenswrapper[4775]: I0321 05:13:21.636935 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6kpp4" event={"ID":"45493002-fd0c-4afd-b556-846d8cac5002","Type":"ContainerStarted","Data":"c4d82c4bdc19edc6f1e2127fa8d601651b0581f9c288d53c6563503a2bf55be8"} Mar 21 05:13:21 crc kubenswrapper[4775]: I0321 05:13:21.660354 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6kpp4" podStartSLOduration=3.870984982 podStartE2EDuration="6.660332906s" podCreationTimestamp="2026-03-21 05:13:15 +0000 UTC" firstStartedPulling="2026-03-21 05:13:17.592171854 +0000 UTC m=+1550.568635478" lastFinishedPulling="2026-03-21 05:13:20.381519778 +0000 UTC m=+1553.357983402" observedRunningTime="2026-03-21 05:13:21.656826248 +0000 UTC m=+1554.633289882" watchObservedRunningTime="2026-03-21 05:13:21.660332906 +0000 UTC m=+1554.636796540" Mar 21 05:13:26 crc kubenswrapper[4775]: I0321 05:13:26.287605 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6kpp4" Mar 21 05:13:26 crc kubenswrapper[4775]: I0321 05:13:26.288081 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6kpp4" Mar 21 05:13:26 crc kubenswrapper[4775]: I0321 05:13:26.336681 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6kpp4" Mar 21 05:13:26 crc kubenswrapper[4775]: I0321 05:13:26.720473 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6kpp4" Mar 21 05:13:26 crc kubenswrapper[4775]: I0321 05:13:26.769105 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6kpp4"] Mar 21 05:13:28 crc kubenswrapper[4775]: I0321 05:13:28.692930 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6kpp4" podUID="45493002-fd0c-4afd-b556-846d8cac5002" containerName="registry-server" containerID="cri-o://c4d82c4bdc19edc6f1e2127fa8d601651b0581f9c288d53c6563503a2bf55be8" gracePeriod=2 Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.182461 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6kpp4" Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.321692 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45493002-fd0c-4afd-b556-846d8cac5002-catalog-content\") pod \"45493002-fd0c-4afd-b556-846d8cac5002\" (UID: \"45493002-fd0c-4afd-b556-846d8cac5002\") " Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.321949 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rglxd\" (UniqueName: \"kubernetes.io/projected/45493002-fd0c-4afd-b556-846d8cac5002-kube-api-access-rglxd\") pod \"45493002-fd0c-4afd-b556-846d8cac5002\" (UID: \"45493002-fd0c-4afd-b556-846d8cac5002\") " Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.322107 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45493002-fd0c-4afd-b556-846d8cac5002-utilities\") pod \"45493002-fd0c-4afd-b556-846d8cac5002\" (UID: \"45493002-fd0c-4afd-b556-846d8cac5002\") " Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.322806 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45493002-fd0c-4afd-b556-846d8cac5002-utilities" (OuterVolumeSpecName: "utilities") pod "45493002-fd0c-4afd-b556-846d8cac5002" (UID: "45493002-fd0c-4afd-b556-846d8cac5002"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.327949 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45493002-fd0c-4afd-b556-846d8cac5002-kube-api-access-rglxd" (OuterVolumeSpecName: "kube-api-access-rglxd") pod "45493002-fd0c-4afd-b556-846d8cac5002" (UID: "45493002-fd0c-4afd-b556-846d8cac5002"). InnerVolumeSpecName "kube-api-access-rglxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.377751 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45493002-fd0c-4afd-b556-846d8cac5002-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45493002-fd0c-4afd-b556-846d8cac5002" (UID: "45493002-fd0c-4afd-b556-846d8cac5002"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.424935 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45493002-fd0c-4afd-b556-846d8cac5002-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.424999 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rglxd\" (UniqueName: \"kubernetes.io/projected/45493002-fd0c-4afd-b556-846d8cac5002-kube-api-access-rglxd\") on node \"crc\" DevicePath \"\"" Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.425070 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45493002-fd0c-4afd-b556-846d8cac5002-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.703145 4775 generic.go:334] "Generic (PLEG): container finished" podID="45493002-fd0c-4afd-b556-846d8cac5002" containerID="c4d82c4bdc19edc6f1e2127fa8d601651b0581f9c288d53c6563503a2bf55be8" exitCode=0 Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.703200 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6kpp4" Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.703199 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6kpp4" event={"ID":"45493002-fd0c-4afd-b556-846d8cac5002","Type":"ContainerDied","Data":"c4d82c4bdc19edc6f1e2127fa8d601651b0581f9c288d53c6563503a2bf55be8"} Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.703339 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6kpp4" event={"ID":"45493002-fd0c-4afd-b556-846d8cac5002","Type":"ContainerDied","Data":"f4eddf91f9b502f94f8ea62789271ca7b116cfd53376cf9077858bb64d6d2b97"} Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.703361 4775 scope.go:117] "RemoveContainer" containerID="c4d82c4bdc19edc6f1e2127fa8d601651b0581f9c288d53c6563503a2bf55be8" Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.725864 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6kpp4"] Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.729430 4775 scope.go:117] "RemoveContainer" containerID="a6782ced417e87e9facb555ea87dc070f4238ca0a19d2f8a047d54825d30f5c4" Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.735629 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6kpp4"] Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.762426 4775 scope.go:117] "RemoveContainer" containerID="cdca173b9c8744ded95289b25e21f20b4cd227c1dd0864ebe8f6347181032830" Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.813445 4775 scope.go:117] "RemoveContainer" containerID="c4d82c4bdc19edc6f1e2127fa8d601651b0581f9c288d53c6563503a2bf55be8" Mar 21 05:13:29 crc kubenswrapper[4775]: E0321 05:13:29.814036 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4d82c4bdc19edc6f1e2127fa8d601651b0581f9c288d53c6563503a2bf55be8\": container with ID starting with c4d82c4bdc19edc6f1e2127fa8d601651b0581f9c288d53c6563503a2bf55be8 not found: ID does not exist" containerID="c4d82c4bdc19edc6f1e2127fa8d601651b0581f9c288d53c6563503a2bf55be8" Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.814067 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4d82c4bdc19edc6f1e2127fa8d601651b0581f9c288d53c6563503a2bf55be8"} err="failed to get container status \"c4d82c4bdc19edc6f1e2127fa8d601651b0581f9c288d53c6563503a2bf55be8\": rpc error: code = NotFound desc = could not find container \"c4d82c4bdc19edc6f1e2127fa8d601651b0581f9c288d53c6563503a2bf55be8\": container with ID starting with c4d82c4bdc19edc6f1e2127fa8d601651b0581f9c288d53c6563503a2bf55be8 not found: ID does not exist" Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.814088 4775 scope.go:117] "RemoveContainer" containerID="a6782ced417e87e9facb555ea87dc070f4238ca0a19d2f8a047d54825d30f5c4" Mar 21 05:13:29 crc kubenswrapper[4775]: E0321 05:13:29.814474 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6782ced417e87e9facb555ea87dc070f4238ca0a19d2f8a047d54825d30f5c4\": container with ID starting with a6782ced417e87e9facb555ea87dc070f4238ca0a19d2f8a047d54825d30f5c4 not found: ID does not exist" containerID="a6782ced417e87e9facb555ea87dc070f4238ca0a19d2f8a047d54825d30f5c4" Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.814529 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6782ced417e87e9facb555ea87dc070f4238ca0a19d2f8a047d54825d30f5c4"} err="failed to get container status \"a6782ced417e87e9facb555ea87dc070f4238ca0a19d2f8a047d54825d30f5c4\": rpc error: code = NotFound desc = could not find container \"a6782ced417e87e9facb555ea87dc070f4238ca0a19d2f8a047d54825d30f5c4\": container with ID starting with a6782ced417e87e9facb555ea87dc070f4238ca0a19d2f8a047d54825d30f5c4 not found: ID does not exist" Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.814562 4775 scope.go:117] "RemoveContainer" containerID="cdca173b9c8744ded95289b25e21f20b4cd227c1dd0864ebe8f6347181032830" Mar 21 05:13:29 crc kubenswrapper[4775]: E0321 05:13:29.815182 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdca173b9c8744ded95289b25e21f20b4cd227c1dd0864ebe8f6347181032830\": container with ID starting with cdca173b9c8744ded95289b25e21f20b4cd227c1dd0864ebe8f6347181032830 not found: ID does not exist" containerID="cdca173b9c8744ded95289b25e21f20b4cd227c1dd0864ebe8f6347181032830" Mar 21 05:13:29 crc kubenswrapper[4775]: I0321 05:13:29.815249 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdca173b9c8744ded95289b25e21f20b4cd227c1dd0864ebe8f6347181032830"} err="failed to get container status \"cdca173b9c8744ded95289b25e21f20b4cd227c1dd0864ebe8f6347181032830\": rpc error: code = NotFound desc = could not find container \"cdca173b9c8744ded95289b25e21f20b4cd227c1dd0864ebe8f6347181032830\": container with ID starting with cdca173b9c8744ded95289b25e21f20b4cd227c1dd0864ebe8f6347181032830 not found: ID does not exist" Mar 21 05:13:31 crc kubenswrapper[4775]: I0321 05:13:31.678303 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45493002-fd0c-4afd-b556-846d8cac5002" path="/var/lib/kubelet/pods/45493002-fd0c-4afd-b556-846d8cac5002/volumes" Mar 21 05:13:32 crc kubenswrapper[4775]: I0321 05:13:32.482770 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:13:32 crc kubenswrapper[4775]: I0321 05:13:32.482835 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:13:44 crc kubenswrapper[4775]: I0321 05:13:44.187886 4775 scope.go:117] "RemoveContainer" containerID="fdcd202238496f57430b4ad2beaac05c4b222263df155d2d3bab8e200bbd88c5" Mar 21 05:14:00 crc kubenswrapper[4775]: I0321 05:14:00.146440 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567834-bqpk6"] Mar 21 05:14:00 crc kubenswrapper[4775]: E0321 05:14:00.147398 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45493002-fd0c-4afd-b556-846d8cac5002" containerName="extract-content" Mar 21 05:14:00 crc kubenswrapper[4775]: I0321 05:14:00.147410 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="45493002-fd0c-4afd-b556-846d8cac5002" containerName="extract-content" Mar 21 05:14:00 crc kubenswrapper[4775]: E0321 05:14:00.147432 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45493002-fd0c-4afd-b556-846d8cac5002" containerName="extract-utilities" Mar 21 05:14:00 crc kubenswrapper[4775]: I0321 05:14:00.147440 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="45493002-fd0c-4afd-b556-846d8cac5002" containerName="extract-utilities" Mar 21 05:14:00 crc kubenswrapper[4775]: E0321 05:14:00.147451 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45493002-fd0c-4afd-b556-846d8cac5002" containerName="registry-server" Mar 21 05:14:00 crc kubenswrapper[4775]: I0321 05:14:00.147458 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="45493002-fd0c-4afd-b556-846d8cac5002" containerName="registry-server" Mar 21 05:14:00 crc kubenswrapper[4775]: I0321 05:14:00.147637 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="45493002-fd0c-4afd-b556-846d8cac5002" containerName="registry-server" Mar 21 05:14:00 crc kubenswrapper[4775]: I0321 05:14:00.148247 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567834-bqpk6" Mar 21 05:14:00 crc kubenswrapper[4775]: I0321 05:14:00.150566 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:14:00 crc kubenswrapper[4775]: I0321 05:14:00.151188 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:14:00 crc kubenswrapper[4775]: I0321 05:14:00.151343 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:14:00 crc kubenswrapper[4775]: I0321 05:14:00.157652 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567834-bqpk6"] Mar 21 05:14:00 crc kubenswrapper[4775]: I0321 05:14:00.332086 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6c2j\" (UniqueName: \"kubernetes.io/projected/dca04e19-ebb7-47c3-a588-e3dc932857bf-kube-api-access-d6c2j\") pod \"auto-csr-approver-29567834-bqpk6\" (UID: \"dca04e19-ebb7-47c3-a588-e3dc932857bf\") " pod="openshift-infra/auto-csr-approver-29567834-bqpk6" Mar 21 05:14:00 crc kubenswrapper[4775]: I0321 05:14:00.433653 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6c2j\" (UniqueName: \"kubernetes.io/projected/dca04e19-ebb7-47c3-a588-e3dc932857bf-kube-api-access-d6c2j\") pod \"auto-csr-approver-29567834-bqpk6\" (UID: \"dca04e19-ebb7-47c3-a588-e3dc932857bf\") " pod="openshift-infra/auto-csr-approver-29567834-bqpk6" Mar 21 05:14:00 crc kubenswrapper[4775]: I0321 05:14:00.451442 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6c2j\" (UniqueName: \"kubernetes.io/projected/dca04e19-ebb7-47c3-a588-e3dc932857bf-kube-api-access-d6c2j\") pod \"auto-csr-approver-29567834-bqpk6\" (UID: \"dca04e19-ebb7-47c3-a588-e3dc932857bf\") " pod="openshift-infra/auto-csr-approver-29567834-bqpk6" Mar 21 05:14:00 crc kubenswrapper[4775]: I0321 05:14:00.466446 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567834-bqpk6" Mar 21 05:14:00 crc kubenswrapper[4775]: I0321 05:14:00.830136 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567834-bqpk6"] Mar 21 05:14:00 crc kubenswrapper[4775]: I0321 05:14:00.852140 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 05:14:00 crc kubenswrapper[4775]: I0321 05:14:00.987904 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567834-bqpk6" event={"ID":"dca04e19-ebb7-47c3-a588-e3dc932857bf","Type":"ContainerStarted","Data":"8ecd5485d0b1d6515fca91ee9e66c8a4fe237e47bea885e8f4c2f9da20eac9d3"} Mar 21 05:14:02 crc kubenswrapper[4775]: I0321 05:14:02.483744 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:14:02 crc kubenswrapper[4775]: I0321 05:14:02.484812 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:14:02 crc kubenswrapper[4775]: I0321 05:14:02.485168 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 05:14:02 crc kubenswrapper[4775]: I0321 05:14:02.486360 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4"} pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 05:14:02 crc kubenswrapper[4775]: I0321 05:14:02.486435 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" containerID="cri-o://5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" gracePeriod=600 Mar 21 05:14:02 crc kubenswrapper[4775]: E0321 05:14:02.633532 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:14:03 crc kubenswrapper[4775]: I0321 05:14:03.010370 4775 generic.go:334] "Generic (PLEG): container finished" podID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" exitCode=0 Mar 21 05:14:03 crc kubenswrapper[4775]: I0321 05:14:03.010433 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerDied","Data":"5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4"} Mar 21 05:14:03 crc kubenswrapper[4775]: I0321 05:14:03.010471 4775 scope.go:117] "RemoveContainer" containerID="b6e85a7c4acc97394b06df812a396284f471ec9c7f9eee22918e9da54e21feda" Mar 21 05:14:03 crc kubenswrapper[4775]: I0321 05:14:03.011540 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:14:03 crc kubenswrapper[4775]: E0321 05:14:03.011974 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:14:04 crc kubenswrapper[4775]: I0321 05:14:04.022733 4775 generic.go:334] "Generic (PLEG): container finished" podID="dca04e19-ebb7-47c3-a588-e3dc932857bf" containerID="a1c552cc0e8c345fb5a90c312288f579e557c98ae97907195961586729f792ec" exitCode=0 Mar 21 05:14:04 crc kubenswrapper[4775]: I0321 05:14:04.022788 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567834-bqpk6" event={"ID":"dca04e19-ebb7-47c3-a588-e3dc932857bf","Type":"ContainerDied","Data":"a1c552cc0e8c345fb5a90c312288f579e557c98ae97907195961586729f792ec"} Mar 21 05:14:05 crc kubenswrapper[4775]: I0321 05:14:05.370087 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567834-bqpk6" Mar 21 05:14:05 crc kubenswrapper[4775]: I0321 05:14:05.443001 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6c2j\" (UniqueName: \"kubernetes.io/projected/dca04e19-ebb7-47c3-a588-e3dc932857bf-kube-api-access-d6c2j\") pod \"dca04e19-ebb7-47c3-a588-e3dc932857bf\" (UID: \"dca04e19-ebb7-47c3-a588-e3dc932857bf\") " Mar 21 05:14:05 crc kubenswrapper[4775]: I0321 05:14:05.448818 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dca04e19-ebb7-47c3-a588-e3dc932857bf-kube-api-access-d6c2j" (OuterVolumeSpecName: "kube-api-access-d6c2j") pod "dca04e19-ebb7-47c3-a588-e3dc932857bf" (UID: "dca04e19-ebb7-47c3-a588-e3dc932857bf"). InnerVolumeSpecName "kube-api-access-d6c2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:14:05 crc kubenswrapper[4775]: I0321 05:14:05.545755 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6c2j\" (UniqueName: \"kubernetes.io/projected/dca04e19-ebb7-47c3-a588-e3dc932857bf-kube-api-access-d6c2j\") on node \"crc\" DevicePath \"\"" Mar 21 05:14:06 crc kubenswrapper[4775]: I0321 05:14:06.047779 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567834-bqpk6" event={"ID":"dca04e19-ebb7-47c3-a588-e3dc932857bf","Type":"ContainerDied","Data":"8ecd5485d0b1d6515fca91ee9e66c8a4fe237e47bea885e8f4c2f9da20eac9d3"} Mar 21 05:14:06 crc kubenswrapper[4775]: I0321 05:14:06.047823 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ecd5485d0b1d6515fca91ee9e66c8a4fe237e47bea885e8f4c2f9da20eac9d3" Mar 21 05:14:06 crc kubenswrapper[4775]: I0321 05:14:06.047837 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567834-bqpk6" Mar 21 05:14:06 crc kubenswrapper[4775]: I0321 05:14:06.437314 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567828-6klnl"] Mar 21 05:14:06 crc kubenswrapper[4775]: I0321 05:14:06.446452 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567828-6klnl"] Mar 21 05:14:07 crc kubenswrapper[4775]: I0321 05:14:07.670781 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eccf102b-f91a-4251-b878-1c21eea92522" path="/var/lib/kubelet/pods/eccf102b-f91a-4251-b878-1c21eea92522/volumes" Mar 21 05:14:17 crc kubenswrapper[4775]: I0321 05:14:17.670766 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:14:17 crc kubenswrapper[4775]: E0321 05:14:17.671958 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:14:31 crc kubenswrapper[4775]: I0321 05:14:31.661595 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:14:31 crc kubenswrapper[4775]: E0321 05:14:31.662515 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:14:44 crc kubenswrapper[4775]: I0321 05:14:44.279058 4775 scope.go:117] "RemoveContainer" containerID="bca1df983777352aa51a410cf20699daccf33fb2338a4057481b7410153afd88" Mar 21 05:14:45 crc kubenswrapper[4775]: I0321 05:14:45.392087 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pjdkb"] Mar 21 05:14:45 crc kubenswrapper[4775]: E0321 05:14:45.392579 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dca04e19-ebb7-47c3-a588-e3dc932857bf" containerName="oc" Mar 21 05:14:45 crc kubenswrapper[4775]: I0321 05:14:45.392594 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="dca04e19-ebb7-47c3-a588-e3dc932857bf" containerName="oc" Mar 21 05:14:45 crc kubenswrapper[4775]: I0321 05:14:45.392853 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="dca04e19-ebb7-47c3-a588-e3dc932857bf" containerName="oc" Mar 21 05:14:45 crc kubenswrapper[4775]: I0321 05:14:45.394499 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pjdkb" Mar 21 05:14:45 crc kubenswrapper[4775]: I0321 05:14:45.419360 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pjdkb"] Mar 21 05:14:45 crc kubenswrapper[4775]: I0321 05:14:45.478275 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20a64964-7adb-48e9-93f6-979650e14968-catalog-content\") pod \"community-operators-pjdkb\" (UID: \"20a64964-7adb-48e9-93f6-979650e14968\") " pod="openshift-marketplace/community-operators-pjdkb" Mar 21 05:14:45 crc kubenswrapper[4775]: I0321 05:14:45.478393 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20a64964-7adb-48e9-93f6-979650e14968-utilities\") pod \"community-operators-pjdkb\" (UID: \"20a64964-7adb-48e9-93f6-979650e14968\") " pod="openshift-marketplace/community-operators-pjdkb" Mar 21 05:14:45 crc kubenswrapper[4775]: I0321 05:14:45.478416 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmlgt\" (UniqueName: \"kubernetes.io/projected/20a64964-7adb-48e9-93f6-979650e14968-kube-api-access-tmlgt\") pod \"community-operators-pjdkb\" (UID: \"20a64964-7adb-48e9-93f6-979650e14968\") " pod="openshift-marketplace/community-operators-pjdkb" Mar 21 05:14:45 crc kubenswrapper[4775]: I0321 05:14:45.579937 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20a64964-7adb-48e9-93f6-979650e14968-catalog-content\") pod \"community-operators-pjdkb\" (UID: \"20a64964-7adb-48e9-93f6-979650e14968\") " pod="openshift-marketplace/community-operators-pjdkb" Mar 21 05:14:45 crc kubenswrapper[4775]: I0321 05:14:45.580320 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20a64964-7adb-48e9-93f6-979650e14968-utilities\") pod \"community-operators-pjdkb\" (UID: \"20a64964-7adb-48e9-93f6-979650e14968\") " pod="openshift-marketplace/community-operators-pjdkb" Mar 21 05:14:45 crc kubenswrapper[4775]: I0321 05:14:45.580352 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmlgt\" (UniqueName: \"kubernetes.io/projected/20a64964-7adb-48e9-93f6-979650e14968-kube-api-access-tmlgt\") pod \"community-operators-pjdkb\" (UID: \"20a64964-7adb-48e9-93f6-979650e14968\") " pod="openshift-marketplace/community-operators-pjdkb" Mar 21 05:14:45 crc kubenswrapper[4775]: I0321 05:14:45.580524 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20a64964-7adb-48e9-93f6-979650e14968-catalog-content\") pod \"community-operators-pjdkb\" (UID: \"20a64964-7adb-48e9-93f6-979650e14968\") " pod="openshift-marketplace/community-operators-pjdkb" Mar 21 05:14:45 crc kubenswrapper[4775]: I0321 05:14:45.580753 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20a64964-7adb-48e9-93f6-979650e14968-utilities\") pod \"community-operators-pjdkb\" (UID: \"20a64964-7adb-48e9-93f6-979650e14968\") " pod="openshift-marketplace/community-operators-pjdkb" Mar 21 05:14:45 crc kubenswrapper[4775]: I0321 05:14:45.605976 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmlgt\" (UniqueName: \"kubernetes.io/projected/20a64964-7adb-48e9-93f6-979650e14968-kube-api-access-tmlgt\") pod \"community-operators-pjdkb\" (UID: \"20a64964-7adb-48e9-93f6-979650e14968\") " pod="openshift-marketplace/community-operators-pjdkb" Mar 21 05:14:45 crc kubenswrapper[4775]: I0321 05:14:45.663820 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:14:45 crc kubenswrapper[4775]: E0321 05:14:45.664028 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:14:45 crc kubenswrapper[4775]: I0321 05:14:45.728703 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pjdkb" Mar 21 05:14:46 crc kubenswrapper[4775]: I0321 05:14:46.251873 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pjdkb"] Mar 21 05:14:46 crc kubenswrapper[4775]: I0321 05:14:46.447734 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjdkb" event={"ID":"20a64964-7adb-48e9-93f6-979650e14968","Type":"ContainerStarted","Data":"7a4739de3c9c5b6e3a614786f2508babd568fd35bee8f59c0d8d1429043ef547"} Mar 21 05:14:47 crc kubenswrapper[4775]: I0321 05:14:47.459243 4775 generic.go:334] "Generic (PLEG): container finished" podID="20a64964-7adb-48e9-93f6-979650e14968" containerID="b7af7cafa1c2a037dbe9f05f5b2616050b352fe6dd4b77bccbb65fd568e4855d" exitCode=0 Mar 21 05:14:47 crc kubenswrapper[4775]: I0321 05:14:47.459338 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjdkb" event={"ID":"20a64964-7adb-48e9-93f6-979650e14968","Type":"ContainerDied","Data":"b7af7cafa1c2a037dbe9f05f5b2616050b352fe6dd4b77bccbb65fd568e4855d"} Mar 21 05:14:48 crc kubenswrapper[4775]: I0321 05:14:48.470559 4775 generic.go:334] "Generic (PLEG): container finished" podID="20a64964-7adb-48e9-93f6-979650e14968" containerID="54f2e589f96425fcbb5095c069f909848ce69dc1edfa225e2b3eaa7471d513e4" exitCode=0 Mar 21 05:14:48 crc kubenswrapper[4775]: I0321 05:14:48.470661 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjdkb" event={"ID":"20a64964-7adb-48e9-93f6-979650e14968","Type":"ContainerDied","Data":"54f2e589f96425fcbb5095c069f909848ce69dc1edfa225e2b3eaa7471d513e4"} Mar 21 05:14:49 crc kubenswrapper[4775]: I0321 05:14:49.481108 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjdkb" event={"ID":"20a64964-7adb-48e9-93f6-979650e14968","Type":"ContainerStarted","Data":"a566da79886c1bc3d992c8c0d8bbadee250657cb273d16a611341ef58c4a8b7d"} Mar 21 05:14:49 crc kubenswrapper[4775]: I0321 05:14:49.497207 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pjdkb" podStartSLOduration=2.986797348 podStartE2EDuration="4.497185421s" podCreationTimestamp="2026-03-21 05:14:45 +0000 UTC" firstStartedPulling="2026-03-21 05:14:47.461700639 +0000 UTC m=+1640.438164263" lastFinishedPulling="2026-03-21 05:14:48.972088712 +0000 UTC m=+1641.948552336" observedRunningTime="2026-03-21 05:14:49.495999407 +0000 UTC m=+1642.472463041" watchObservedRunningTime="2026-03-21 05:14:49.497185421 +0000 UTC m=+1642.473649045" Mar 21 05:14:55 crc kubenswrapper[4775]: I0321 05:14:55.729802 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pjdkb" Mar 21 05:14:55 crc kubenswrapper[4775]: I0321 05:14:55.730208 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pjdkb" Mar 21 05:14:55 crc kubenswrapper[4775]: I0321 05:14:55.776575 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pjdkb" Mar 21 05:14:56 crc kubenswrapper[4775]: I0321 05:14:56.581558 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pjdkb" Mar 21 05:14:56 crc kubenswrapper[4775]: I0321 05:14:56.649803 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pjdkb"] Mar 21 05:14:56 crc kubenswrapper[4775]: I0321 05:14:56.661564 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:14:56 crc kubenswrapper[4775]: E0321 05:14:56.661828 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:14:58 crc kubenswrapper[4775]: I0321 05:14:58.558987 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pjdkb" podUID="20a64964-7adb-48e9-93f6-979650e14968" containerName="registry-server" containerID="cri-o://a566da79886c1bc3d992c8c0d8bbadee250657cb273d16a611341ef58c4a8b7d" gracePeriod=2 Mar 21 05:14:58 crc kubenswrapper[4775]: I0321 05:14:58.958802 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pjdkb" Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.152336 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20a64964-7adb-48e9-93f6-979650e14968-catalog-content\") pod \"20a64964-7adb-48e9-93f6-979650e14968\" (UID: \"20a64964-7adb-48e9-93f6-979650e14968\") " Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.152518 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20a64964-7adb-48e9-93f6-979650e14968-utilities\") pod \"20a64964-7adb-48e9-93f6-979650e14968\" (UID: \"20a64964-7adb-48e9-93f6-979650e14968\") " Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.152600 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmlgt\" (UniqueName: \"kubernetes.io/projected/20a64964-7adb-48e9-93f6-979650e14968-kube-api-access-tmlgt\") pod \"20a64964-7adb-48e9-93f6-979650e14968\" (UID: \"20a64964-7adb-48e9-93f6-979650e14968\") " Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.153205 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20a64964-7adb-48e9-93f6-979650e14968-utilities" (OuterVolumeSpecName: "utilities") pod "20a64964-7adb-48e9-93f6-979650e14968" (UID: "20a64964-7adb-48e9-93f6-979650e14968"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.164021 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20a64964-7adb-48e9-93f6-979650e14968-kube-api-access-tmlgt" (OuterVolumeSpecName: "kube-api-access-tmlgt") pod "20a64964-7adb-48e9-93f6-979650e14968" (UID: "20a64964-7adb-48e9-93f6-979650e14968"). InnerVolumeSpecName "kube-api-access-tmlgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.209782 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20a64964-7adb-48e9-93f6-979650e14968-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20a64964-7adb-48e9-93f6-979650e14968" (UID: "20a64964-7adb-48e9-93f6-979650e14968"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.255467 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20a64964-7adb-48e9-93f6-979650e14968-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.255502 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20a64964-7adb-48e9-93f6-979650e14968-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.255512 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmlgt\" (UniqueName: \"kubernetes.io/projected/20a64964-7adb-48e9-93f6-979650e14968-kube-api-access-tmlgt\") on node \"crc\" DevicePath \"\"" Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.573938 4775 generic.go:334] "Generic (PLEG): container finished" podID="20a64964-7adb-48e9-93f6-979650e14968" containerID="a566da79886c1bc3d992c8c0d8bbadee250657cb273d16a611341ef58c4a8b7d" exitCode=0 Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.573991 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjdkb" event={"ID":"20a64964-7adb-48e9-93f6-979650e14968","Type":"ContainerDied","Data":"a566da79886c1bc3d992c8c0d8bbadee250657cb273d16a611341ef58c4a8b7d"} Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.574027 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjdkb" event={"ID":"20a64964-7adb-48e9-93f6-979650e14968","Type":"ContainerDied","Data":"7a4739de3c9c5b6e3a614786f2508babd568fd35bee8f59c0d8d1429043ef547"} Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.574046 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pjdkb" Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.574054 4775 scope.go:117] "RemoveContainer" containerID="a566da79886c1bc3d992c8c0d8bbadee250657cb273d16a611341ef58c4a8b7d" Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.617695 4775 scope.go:117] "RemoveContainer" containerID="54f2e589f96425fcbb5095c069f909848ce69dc1edfa225e2b3eaa7471d513e4" Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.620094 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pjdkb"] Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.629522 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pjdkb"] Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.649799 4775 scope.go:117] "RemoveContainer" containerID="b7af7cafa1c2a037dbe9f05f5b2616050b352fe6dd4b77bccbb65fd568e4855d" Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.674071 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20a64964-7adb-48e9-93f6-979650e14968" path="/var/lib/kubelet/pods/20a64964-7adb-48e9-93f6-979650e14968/volumes" Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.693379 4775 scope.go:117] "RemoveContainer" containerID="a566da79886c1bc3d992c8c0d8bbadee250657cb273d16a611341ef58c4a8b7d" Mar 21 05:14:59 crc kubenswrapper[4775]: E0321 05:14:59.693816 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a566da79886c1bc3d992c8c0d8bbadee250657cb273d16a611341ef58c4a8b7d\": container with ID starting with a566da79886c1bc3d992c8c0d8bbadee250657cb273d16a611341ef58c4a8b7d not found: ID does not exist" containerID="a566da79886c1bc3d992c8c0d8bbadee250657cb273d16a611341ef58c4a8b7d" Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.693871 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a566da79886c1bc3d992c8c0d8bbadee250657cb273d16a611341ef58c4a8b7d"} err="failed to get container status \"a566da79886c1bc3d992c8c0d8bbadee250657cb273d16a611341ef58c4a8b7d\": rpc error: code = NotFound desc = could not find container \"a566da79886c1bc3d992c8c0d8bbadee250657cb273d16a611341ef58c4a8b7d\": container with ID starting with a566da79886c1bc3d992c8c0d8bbadee250657cb273d16a611341ef58c4a8b7d not found: ID does not exist" Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.693920 4775 scope.go:117] "RemoveContainer" containerID="54f2e589f96425fcbb5095c069f909848ce69dc1edfa225e2b3eaa7471d513e4" Mar 21 05:14:59 crc kubenswrapper[4775]: E0321 05:14:59.694341 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54f2e589f96425fcbb5095c069f909848ce69dc1edfa225e2b3eaa7471d513e4\": container with ID starting with 54f2e589f96425fcbb5095c069f909848ce69dc1edfa225e2b3eaa7471d513e4 not found: ID does not exist" containerID="54f2e589f96425fcbb5095c069f909848ce69dc1edfa225e2b3eaa7471d513e4" Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.694370 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54f2e589f96425fcbb5095c069f909848ce69dc1edfa225e2b3eaa7471d513e4"} err="failed to get container status \"54f2e589f96425fcbb5095c069f909848ce69dc1edfa225e2b3eaa7471d513e4\": rpc error: code = NotFound desc = could not find container \"54f2e589f96425fcbb5095c069f909848ce69dc1edfa225e2b3eaa7471d513e4\": container with ID starting with 54f2e589f96425fcbb5095c069f909848ce69dc1edfa225e2b3eaa7471d513e4 not found: ID does not exist" Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.694392 4775 scope.go:117] "RemoveContainer" containerID="b7af7cafa1c2a037dbe9f05f5b2616050b352fe6dd4b77bccbb65fd568e4855d" Mar 21 05:14:59 crc kubenswrapper[4775]: E0321 05:14:59.694805 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7af7cafa1c2a037dbe9f05f5b2616050b352fe6dd4b77bccbb65fd568e4855d\": container with ID starting with b7af7cafa1c2a037dbe9f05f5b2616050b352fe6dd4b77bccbb65fd568e4855d not found: ID does not exist" containerID="b7af7cafa1c2a037dbe9f05f5b2616050b352fe6dd4b77bccbb65fd568e4855d" Mar 21 05:14:59 crc kubenswrapper[4775]: I0321 05:14:59.694835 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7af7cafa1c2a037dbe9f05f5b2616050b352fe6dd4b77bccbb65fd568e4855d"} err="failed to get container status \"b7af7cafa1c2a037dbe9f05f5b2616050b352fe6dd4b77bccbb65fd568e4855d\": rpc error: code = NotFound desc = could not find container \"b7af7cafa1c2a037dbe9f05f5b2616050b352fe6dd4b77bccbb65fd568e4855d\": container with ID starting with b7af7cafa1c2a037dbe9f05f5b2616050b352fe6dd4b77bccbb65fd568e4855d not found: ID does not exist" Mar 21 05:15:00 crc kubenswrapper[4775]: I0321 05:15:00.151829 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx"] Mar 21 05:15:00 crc kubenswrapper[4775]: E0321 05:15:00.152457 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a64964-7adb-48e9-93f6-979650e14968" containerName="registry-server" Mar 21 05:15:00 crc kubenswrapper[4775]: I0321 05:15:00.152486 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a64964-7adb-48e9-93f6-979650e14968" containerName="registry-server" Mar 21 05:15:00 crc kubenswrapper[4775]: E0321 05:15:00.152512 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a64964-7adb-48e9-93f6-979650e14968" containerName="extract-content" Mar 21 05:15:00 crc kubenswrapper[4775]: I0321 05:15:00.152520 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a64964-7adb-48e9-93f6-979650e14968" containerName="extract-content" Mar 21 05:15:00 crc kubenswrapper[4775]: E0321 05:15:00.152535 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a64964-7adb-48e9-93f6-979650e14968" containerName="extract-utilities" Mar 21 05:15:00 crc kubenswrapper[4775]: I0321 05:15:00.152543 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a64964-7adb-48e9-93f6-979650e14968" containerName="extract-utilities" Mar 21 05:15:00 crc kubenswrapper[4775]: I0321 05:15:00.152987 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="20a64964-7adb-48e9-93f6-979650e14968" containerName="registry-server" Mar 21 05:15:00 crc kubenswrapper[4775]: I0321 05:15:00.154658 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx" Mar 21 05:15:00 crc kubenswrapper[4775]: I0321 05:15:00.160771 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx"] Mar 21 05:15:00 crc kubenswrapper[4775]: I0321 05:15:00.161655 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 21 05:15:00 crc kubenswrapper[4775]: I0321 05:15:00.161954 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 21 05:15:00 crc kubenswrapper[4775]: I0321 05:15:00.275719 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bebab26-7b09-4e00-adee-fffa3c06f5ab-secret-volume\") pod \"collect-profiles-29567835-47dnx\" (UID: \"8bebab26-7b09-4e00-adee-fffa3c06f5ab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx" Mar 21 05:15:00 crc kubenswrapper[4775]: I0321 05:15:00.276147 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bebab26-7b09-4e00-adee-fffa3c06f5ab-config-volume\") pod \"collect-profiles-29567835-47dnx\" (UID: \"8bebab26-7b09-4e00-adee-fffa3c06f5ab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx" Mar 21 05:15:00 crc kubenswrapper[4775]: I0321 05:15:00.276252 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kb749\" (UniqueName: \"kubernetes.io/projected/8bebab26-7b09-4e00-adee-fffa3c06f5ab-kube-api-access-kb749\") pod \"collect-profiles-29567835-47dnx\" (UID: \"8bebab26-7b09-4e00-adee-fffa3c06f5ab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx" Mar 21 05:15:00 crc kubenswrapper[4775]: I0321 05:15:00.378384 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bebab26-7b09-4e00-adee-fffa3c06f5ab-config-volume\") pod \"collect-profiles-29567835-47dnx\" (UID: \"8bebab26-7b09-4e00-adee-fffa3c06f5ab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx" Mar 21 05:15:00 crc kubenswrapper[4775]: I0321 05:15:00.378514 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kb749\" (UniqueName: \"kubernetes.io/projected/8bebab26-7b09-4e00-adee-fffa3c06f5ab-kube-api-access-kb749\") pod \"collect-profiles-29567835-47dnx\" (UID: \"8bebab26-7b09-4e00-adee-fffa3c06f5ab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx" Mar 21 05:15:00 crc kubenswrapper[4775]: I0321 05:15:00.378570 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bebab26-7b09-4e00-adee-fffa3c06f5ab-secret-volume\") pod \"collect-profiles-29567835-47dnx\" (UID: \"8bebab26-7b09-4e00-adee-fffa3c06f5ab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx" Mar 21 05:15:00 crc kubenswrapper[4775]: I0321 05:15:00.379380 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bebab26-7b09-4e00-adee-fffa3c06f5ab-config-volume\") pod \"collect-profiles-29567835-47dnx\" (UID: \"8bebab26-7b09-4e00-adee-fffa3c06f5ab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx" Mar 21 05:15:00 crc kubenswrapper[4775]: I0321 05:15:00.384927 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bebab26-7b09-4e00-adee-fffa3c06f5ab-secret-volume\") pod \"collect-profiles-29567835-47dnx\" (UID: \"8bebab26-7b09-4e00-adee-fffa3c06f5ab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx" Mar 21 05:15:00 crc kubenswrapper[4775]: I0321 05:15:00.399110 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kb749\" (UniqueName: \"kubernetes.io/projected/8bebab26-7b09-4e00-adee-fffa3c06f5ab-kube-api-access-kb749\") pod \"collect-profiles-29567835-47dnx\" (UID: \"8bebab26-7b09-4e00-adee-fffa3c06f5ab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx" Mar 21 05:15:00 crc kubenswrapper[4775]: I0321 05:15:00.489932 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx" Mar 21 05:15:00 crc kubenswrapper[4775]: I0321 05:15:00.949798 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx"] Mar 21 05:15:01 crc kubenswrapper[4775]: I0321 05:15:01.623675 4775 generic.go:334] "Generic (PLEG): container finished" podID="8bebab26-7b09-4e00-adee-fffa3c06f5ab" containerID="c3748d6274f12563ae1762dcd4c9c668e2bf7b8aecb55d574f50d517af661d16" exitCode=0 Mar 21 05:15:01 crc kubenswrapper[4775]: I0321 05:15:01.623755 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx" event={"ID":"8bebab26-7b09-4e00-adee-fffa3c06f5ab","Type":"ContainerDied","Data":"c3748d6274f12563ae1762dcd4c9c668e2bf7b8aecb55d574f50d517af661d16"} Mar 21 05:15:01 crc kubenswrapper[4775]: I0321 05:15:01.623946 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx" event={"ID":"8bebab26-7b09-4e00-adee-fffa3c06f5ab","Type":"ContainerStarted","Data":"4e852416a815e8ad2197d4a13c6cdcc4844025c71374e05497715f2c9a4e7420"} Mar 21 05:15:01 crc kubenswrapper[4775]: I0321 05:15:01.626818 4775 generic.go:334] "Generic (PLEG): container finished" podID="203df932-0574-4098-b897-ba50813f2ec1" containerID="e83db7ccdc230f6a7c372f2ab3e28ef862efd0646a039a1d8f53714e19540048" exitCode=0 Mar 21 05:15:01 crc kubenswrapper[4775]: I0321 05:15:01.626870 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" event={"ID":"203df932-0574-4098-b897-ba50813f2ec1","Type":"ContainerDied","Data":"e83db7ccdc230f6a7c372f2ab3e28ef862efd0646a039a1d8f53714e19540048"} Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.051798 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.059069 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.138096 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sfxh\" (UniqueName: \"kubernetes.io/projected/203df932-0574-4098-b897-ba50813f2ec1-kube-api-access-5sfxh\") pod \"203df932-0574-4098-b897-ba50813f2ec1\" (UID: \"203df932-0574-4098-b897-ba50813f2ec1\") " Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.138216 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kb749\" (UniqueName: \"kubernetes.io/projected/8bebab26-7b09-4e00-adee-fffa3c06f5ab-kube-api-access-kb749\") pod \"8bebab26-7b09-4e00-adee-fffa3c06f5ab\" (UID: \"8bebab26-7b09-4e00-adee-fffa3c06f5ab\") " Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.138305 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bebab26-7b09-4e00-adee-fffa3c06f5ab-secret-volume\") pod \"8bebab26-7b09-4e00-adee-fffa3c06f5ab\" (UID: \"8bebab26-7b09-4e00-adee-fffa3c06f5ab\") " Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.138526 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bebab26-7b09-4e00-adee-fffa3c06f5ab-config-volume\") pod \"8bebab26-7b09-4e00-adee-fffa3c06f5ab\" (UID: \"8bebab26-7b09-4e00-adee-fffa3c06f5ab\") " Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.138624 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/203df932-0574-4098-b897-ba50813f2ec1-inventory\") pod \"203df932-0574-4098-b897-ba50813f2ec1\" (UID: \"203df932-0574-4098-b897-ba50813f2ec1\") " Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.138739 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/203df932-0574-4098-b897-ba50813f2ec1-bootstrap-combined-ca-bundle\") pod \"203df932-0574-4098-b897-ba50813f2ec1\" (UID: \"203df932-0574-4098-b897-ba50813f2ec1\") " Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.138778 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/203df932-0574-4098-b897-ba50813f2ec1-ssh-key-openstack-edpm-ipam\") pod \"203df932-0574-4098-b897-ba50813f2ec1\" (UID: \"203df932-0574-4098-b897-ba50813f2ec1\") " Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.139605 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bebab26-7b09-4e00-adee-fffa3c06f5ab-config-volume" (OuterVolumeSpecName: "config-volume") pod "8bebab26-7b09-4e00-adee-fffa3c06f5ab" (UID: "8bebab26-7b09-4e00-adee-fffa3c06f5ab"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.144716 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bebab26-7b09-4e00-adee-fffa3c06f5ab-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8bebab26-7b09-4e00-adee-fffa3c06f5ab" (UID: "8bebab26-7b09-4e00-adee-fffa3c06f5ab"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.145004 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bebab26-7b09-4e00-adee-fffa3c06f5ab-kube-api-access-kb749" (OuterVolumeSpecName: "kube-api-access-kb749") pod "8bebab26-7b09-4e00-adee-fffa3c06f5ab" (UID: "8bebab26-7b09-4e00-adee-fffa3c06f5ab"). InnerVolumeSpecName "kube-api-access-kb749". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.145058 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/203df932-0574-4098-b897-ba50813f2ec1-kube-api-access-5sfxh" (OuterVolumeSpecName: "kube-api-access-5sfxh") pod "203df932-0574-4098-b897-ba50813f2ec1" (UID: "203df932-0574-4098-b897-ba50813f2ec1"). InnerVolumeSpecName "kube-api-access-5sfxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.145509 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/203df932-0574-4098-b897-ba50813f2ec1-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "203df932-0574-4098-b897-ba50813f2ec1" (UID: "203df932-0574-4098-b897-ba50813f2ec1"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.168835 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/203df932-0574-4098-b897-ba50813f2ec1-inventory" (OuterVolumeSpecName: "inventory") pod "203df932-0574-4098-b897-ba50813f2ec1" (UID: "203df932-0574-4098-b897-ba50813f2ec1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.182094 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/203df932-0574-4098-b897-ba50813f2ec1-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "203df932-0574-4098-b897-ba50813f2ec1" (UID: "203df932-0574-4098-b897-ba50813f2ec1"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.240607 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/203df932-0574-4098-b897-ba50813f2ec1-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.240646 4775 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/203df932-0574-4098-b897-ba50813f2ec1-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.240657 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/203df932-0574-4098-b897-ba50813f2ec1-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.240667 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sfxh\" (UniqueName: \"kubernetes.io/projected/203df932-0574-4098-b897-ba50813f2ec1-kube-api-access-5sfxh\") on node \"crc\" DevicePath \"\"" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.240676 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kb749\" (UniqueName: \"kubernetes.io/projected/8bebab26-7b09-4e00-adee-fffa3c06f5ab-kube-api-access-kb749\") on node \"crc\" DevicePath \"\"" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.240684 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bebab26-7b09-4e00-adee-fffa3c06f5ab-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.240693 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bebab26-7b09-4e00-adee-fffa3c06f5ab-config-volume\") on node \"crc\" DevicePath \"\"" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.648650 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx" event={"ID":"8bebab26-7b09-4e00-adee-fffa3c06f5ab","Type":"ContainerDied","Data":"4e852416a815e8ad2197d4a13c6cdcc4844025c71374e05497715f2c9a4e7420"} Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.648982 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e852416a815e8ad2197d4a13c6cdcc4844025c71374e05497715f2c9a4e7420" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.648672 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.650600 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" event={"ID":"203df932-0574-4098-b897-ba50813f2ec1","Type":"ContainerDied","Data":"3c801bc6a121fd4faebee72f398b0bcdeabcebcace5ab4c44442f937b77950dc"} Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.650639 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c801bc6a121fd4faebee72f398b0bcdeabcebcace5ab4c44442f937b77950dc" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.650680 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.753338 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr"] Mar 21 05:15:03 crc kubenswrapper[4775]: E0321 05:15:03.753832 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="203df932-0574-4098-b897-ba50813f2ec1" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.753857 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="203df932-0574-4098-b897-ba50813f2ec1" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 21 05:15:03 crc kubenswrapper[4775]: E0321 05:15:03.753900 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bebab26-7b09-4e00-adee-fffa3c06f5ab" containerName="collect-profiles" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.753914 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bebab26-7b09-4e00-adee-fffa3c06f5ab" containerName="collect-profiles" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.754155 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bebab26-7b09-4e00-adee-fffa3c06f5ab" containerName="collect-profiles" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.754190 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="203df932-0574-4098-b897-ba50813f2ec1" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.755010 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.757516 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.757815 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljtt5" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.758439 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.758699 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.771867 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr"] Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.850663 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/28040d61-c9ea-4a55-b113-db871dff679c-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-h87nr\" (UID: \"28040d61-c9ea-4a55-b113-db871dff679c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.850794 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28040d61-c9ea-4a55-b113-db871dff679c-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-h87nr\" (UID: \"28040d61-c9ea-4a55-b113-db871dff679c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.850891 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ps8td\" (UniqueName: \"kubernetes.io/projected/28040d61-c9ea-4a55-b113-db871dff679c-kube-api-access-ps8td\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-h87nr\" (UID: \"28040d61-c9ea-4a55-b113-db871dff679c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.956110 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/28040d61-c9ea-4a55-b113-db871dff679c-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-h87nr\" (UID: \"28040d61-c9ea-4a55-b113-db871dff679c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.956397 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28040d61-c9ea-4a55-b113-db871dff679c-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-h87nr\" (UID: \"28040d61-c9ea-4a55-b113-db871dff679c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.956562 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ps8td\" (UniqueName: \"kubernetes.io/projected/28040d61-c9ea-4a55-b113-db871dff679c-kube-api-access-ps8td\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-h87nr\" (UID: \"28040d61-c9ea-4a55-b113-db871dff679c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.962553 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/28040d61-c9ea-4a55-b113-db871dff679c-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-h87nr\" (UID: \"28040d61-c9ea-4a55-b113-db871dff679c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.966768 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28040d61-c9ea-4a55-b113-db871dff679c-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-h87nr\" (UID: \"28040d61-c9ea-4a55-b113-db871dff679c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr" Mar 21 05:15:03 crc kubenswrapper[4775]: I0321 05:15:03.984083 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ps8td\" (UniqueName: \"kubernetes.io/projected/28040d61-c9ea-4a55-b113-db871dff679c-kube-api-access-ps8td\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-h87nr\" (UID: \"28040d61-c9ea-4a55-b113-db871dff679c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr" Mar 21 05:15:04 crc kubenswrapper[4775]: I0321 05:15:04.089432 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr" Mar 21 05:15:04 crc kubenswrapper[4775]: I0321 05:15:04.630552 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr"] Mar 21 05:15:04 crc kubenswrapper[4775]: I0321 05:15:04.659616 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr" event={"ID":"28040d61-c9ea-4a55-b113-db871dff679c","Type":"ContainerStarted","Data":"ef0fcdf856bb63344889100ab23ff6e3c752ecddd8029d7da572bb02105b88a6"} Mar 21 05:15:06 crc kubenswrapper[4775]: I0321 05:15:06.680889 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr" event={"ID":"28040d61-c9ea-4a55-b113-db871dff679c","Type":"ContainerStarted","Data":"4f6b82cb9ea1aa1f200cad1ffb7a707dc8d8464ef48ddbc1ece993160bc193ff"} Mar 21 05:15:06 crc kubenswrapper[4775]: I0321 05:15:06.702518 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr" podStartSLOduration=2.357774997 podStartE2EDuration="3.702497304s" podCreationTimestamp="2026-03-21 05:15:03 +0000 UTC" firstStartedPulling="2026-03-21 05:15:04.632323881 +0000 UTC m=+1657.608787505" lastFinishedPulling="2026-03-21 05:15:05.977046178 +0000 UTC m=+1658.953509812" observedRunningTime="2026-03-21 05:15:06.692642308 +0000 UTC m=+1659.669105932" watchObservedRunningTime="2026-03-21 05:15:06.702497304 +0000 UTC m=+1659.678960928" Mar 21 05:15:10 crc kubenswrapper[4775]: I0321 05:15:10.661555 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:15:10 crc kubenswrapper[4775]: E0321 05:15:10.662420 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:15:22 crc kubenswrapper[4775]: I0321 05:15:22.661423 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:15:22 crc kubenswrapper[4775]: E0321 05:15:22.662136 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:15:37 crc kubenswrapper[4775]: I0321 05:15:37.671343 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:15:37 crc kubenswrapper[4775]: E0321 05:15:37.672160 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:15:44 crc kubenswrapper[4775]: I0321 05:15:44.383779 4775 scope.go:117] "RemoveContainer" containerID="3d41a9111ff02a22981deef2e158c7aaae04c761a3f477f9cde2ecff8054c0f1" Mar 21 05:15:44 crc kubenswrapper[4775]: I0321 05:15:44.421986 4775 scope.go:117] "RemoveContainer" containerID="992670c46bdc3ecfcf02d509f68efcc94de86d4d365af73eee5b378735e764a6" Mar 21 05:15:44 crc kubenswrapper[4775]: I0321 05:15:44.451778 4775 scope.go:117] "RemoveContainer" containerID="a3e8af6598015c259cd8e37d94c33e2e14ca3a07bdbbfc459894647fdb9e40ea" Mar 21 05:15:44 crc kubenswrapper[4775]: I0321 05:15:44.515073 4775 scope.go:117] "RemoveContainer" containerID="210c52f72b30e6d06b1636b7f4ba7c385a206b219bca52c176d9da81b984a93a" Mar 21 05:15:48 crc kubenswrapper[4775]: I0321 05:15:48.662674 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:15:48 crc kubenswrapper[4775]: E0321 05:15:48.663565 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:16:00 crc kubenswrapper[4775]: I0321 05:16:00.168299 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567836-whhlh"] Mar 21 05:16:00 crc kubenswrapper[4775]: I0321 05:16:00.170876 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567836-whhlh" Mar 21 05:16:00 crc kubenswrapper[4775]: I0321 05:16:00.174547 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:16:00 crc kubenswrapper[4775]: I0321 05:16:00.177750 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:16:00 crc kubenswrapper[4775]: I0321 05:16:00.177854 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:16:00 crc kubenswrapper[4775]: I0321 05:16:00.188203 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567836-whhlh"] Mar 21 05:16:00 crc kubenswrapper[4775]: I0321 05:16:00.329436 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvlxw\" (UniqueName: \"kubernetes.io/projected/4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc-kube-api-access-hvlxw\") pod \"auto-csr-approver-29567836-whhlh\" (UID: \"4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc\") " pod="openshift-infra/auto-csr-approver-29567836-whhlh" Mar 21 05:16:00 crc kubenswrapper[4775]: I0321 05:16:00.431795 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvlxw\" (UniqueName: \"kubernetes.io/projected/4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc-kube-api-access-hvlxw\") pod \"auto-csr-approver-29567836-whhlh\" (UID: \"4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc\") " pod="openshift-infra/auto-csr-approver-29567836-whhlh" Mar 21 05:16:00 crc kubenswrapper[4775]: I0321 05:16:00.455567 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvlxw\" (UniqueName: \"kubernetes.io/projected/4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc-kube-api-access-hvlxw\") pod \"auto-csr-approver-29567836-whhlh\" (UID: \"4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc\") " pod="openshift-infra/auto-csr-approver-29567836-whhlh" Mar 21 05:16:00 crc kubenswrapper[4775]: I0321 05:16:00.497850 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567836-whhlh" Mar 21 05:16:01 crc kubenswrapper[4775]: I0321 05:16:01.068896 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567836-whhlh"] Mar 21 05:16:01 crc kubenswrapper[4775]: I0321 05:16:01.257585 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567836-whhlh" event={"ID":"4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc","Type":"ContainerStarted","Data":"f61d14db2b5b45ca660eea6cc2d22453b0890440d252c939cb8c03e9cda1c42a"} Mar 21 05:16:01 crc kubenswrapper[4775]: I0321 05:16:01.661897 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:16:01 crc kubenswrapper[4775]: E0321 05:16:01.662343 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:16:04 crc kubenswrapper[4775]: I0321 05:16:04.291725 4775 generic.go:334] "Generic (PLEG): container finished" podID="4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc" containerID="f116b6d526cd4aaa616f201bcf5084943e1e56754ce61bd36026bd8732dd3bd1" exitCode=0 Mar 21 05:16:04 crc kubenswrapper[4775]: I0321 05:16:04.291783 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567836-whhlh" event={"ID":"4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc","Type":"ContainerDied","Data":"f116b6d526cd4aaa616f201bcf5084943e1e56754ce61bd36026bd8732dd3bd1"} Mar 21 05:16:05 crc kubenswrapper[4775]: I0321 05:16:05.830825 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567836-whhlh" Mar 21 05:16:05 crc kubenswrapper[4775]: I0321 05:16:05.904834 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvlxw\" (UniqueName: \"kubernetes.io/projected/4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc-kube-api-access-hvlxw\") pod \"4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc\" (UID: \"4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc\") " Mar 21 05:16:05 crc kubenswrapper[4775]: I0321 05:16:05.919211 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc-kube-api-access-hvlxw" (OuterVolumeSpecName: "kube-api-access-hvlxw") pod "4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc" (UID: "4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc"). InnerVolumeSpecName "kube-api-access-hvlxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:16:06 crc kubenswrapper[4775]: I0321 05:16:06.009077 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvlxw\" (UniqueName: \"kubernetes.io/projected/4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc-kube-api-access-hvlxw\") on node \"crc\" DevicePath \"\"" Mar 21 05:16:06 crc kubenswrapper[4775]: I0321 05:16:06.314790 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567836-whhlh" event={"ID":"4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc","Type":"ContainerDied","Data":"f61d14db2b5b45ca660eea6cc2d22453b0890440d252c939cb8c03e9cda1c42a"} Mar 21 05:16:06 crc kubenswrapper[4775]: I0321 05:16:06.315092 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f61d14db2b5b45ca660eea6cc2d22453b0890440d252c939cb8c03e9cda1c42a" Mar 21 05:16:06 crc kubenswrapper[4775]: I0321 05:16:06.314952 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567836-whhlh" Mar 21 05:16:06 crc kubenswrapper[4775]: I0321 05:16:06.921417 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567830-jv78x"] Mar 21 05:16:06 crc kubenswrapper[4775]: I0321 05:16:06.933669 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567830-jv78x"] Mar 21 05:16:07 crc kubenswrapper[4775]: I0321 05:16:07.678946 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33cb4f91-aca0-4221-873f-e2c16e30ccee" path="/var/lib/kubelet/pods/33cb4f91-aca0-4221-873f-e2c16e30ccee/volumes" Mar 21 05:16:11 crc kubenswrapper[4775]: I0321 05:16:11.042545 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-fmfms"] Mar 21 05:16:11 crc kubenswrapper[4775]: I0321 05:16:11.051574 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-fmfms"] Mar 21 05:16:11 crc kubenswrapper[4775]: I0321 05:16:11.675503 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a151b65-5187-48d2-a29a-30eae42c179c" path="/var/lib/kubelet/pods/3a151b65-5187-48d2-a29a-30eae42c179c/volumes" Mar 21 05:16:12 crc kubenswrapper[4775]: I0321 05:16:12.054094 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-6de1-account-create-update-bkzxl"] Mar 21 05:16:12 crc kubenswrapper[4775]: I0321 05:16:12.067408 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-594f-account-create-update-9ntmw"] Mar 21 05:16:12 crc kubenswrapper[4775]: I0321 05:16:12.078999 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-6de1-account-create-update-bkzxl"] Mar 21 05:16:12 crc kubenswrapper[4775]: I0321 05:16:12.090804 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-429fj"] Mar 21 05:16:12 crc kubenswrapper[4775]: I0321 05:16:12.102598 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-594f-account-create-update-9ntmw"] Mar 21 05:16:12 crc kubenswrapper[4775]: I0321 05:16:12.113661 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-429fj"] Mar 21 05:16:13 crc kubenswrapper[4775]: I0321 05:16:13.042535 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-41dd-account-create-update-5ltn6"] Mar 21 05:16:13 crc kubenswrapper[4775]: I0321 05:16:13.058039 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-j56nc"] Mar 21 05:16:13 crc kubenswrapper[4775]: I0321 05:16:13.071847 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-j56nc"] Mar 21 05:16:13 crc kubenswrapper[4775]: I0321 05:16:13.083326 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-41dd-account-create-update-5ltn6"] Mar 21 05:16:13 crc kubenswrapper[4775]: I0321 05:16:13.675912 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6" path="/var/lib/kubelet/pods/1623bfa5-c196-4b3f-b1ad-2e895cb6e6d6/volumes" Mar 21 05:16:13 crc kubenswrapper[4775]: I0321 05:16:13.676684 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17e15fc9-dccb-44f6-9266-2bfe23d9e224" path="/var/lib/kubelet/pods/17e15fc9-dccb-44f6-9266-2bfe23d9e224/volumes" Mar 21 05:16:13 crc kubenswrapper[4775]: I0321 05:16:13.677746 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88db3052-d0c3-4f00-a116-aea162a7790b" path="/var/lib/kubelet/pods/88db3052-d0c3-4f00-a116-aea162a7790b/volumes" Mar 21 05:16:13 crc kubenswrapper[4775]: I0321 05:16:13.678513 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5ecd8b9-b671-4181-96e9-adf3e820c8c7" path="/var/lib/kubelet/pods/a5ecd8b9-b671-4181-96e9-adf3e820c8c7/volumes" Mar 21 05:16:13 crc kubenswrapper[4775]: I0321 05:16:13.679308 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8c83c73-8982-4c8e-bdeb-204a4874162c" path="/var/lib/kubelet/pods/a8c83c73-8982-4c8e-bdeb-204a4874162c/volumes" Mar 21 05:16:14 crc kubenswrapper[4775]: I0321 05:16:14.662832 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:16:14 crc kubenswrapper[4775]: E0321 05:16:14.663697 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:16:25 crc kubenswrapper[4775]: I0321 05:16:25.663557 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:16:25 crc kubenswrapper[4775]: E0321 05:16:25.664839 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:16:32 crc kubenswrapper[4775]: I0321 05:16:32.045914 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-drcfj"] Mar 21 05:16:32 crc kubenswrapper[4775]: I0321 05:16:32.059756 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-drcfj"] Mar 21 05:16:33 crc kubenswrapper[4775]: I0321 05:16:33.676173 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4c6072b-3aa5-43ef-be24-f9b20e5095bd" path="/var/lib/kubelet/pods/b4c6072b-3aa5-43ef-be24-f9b20e5095bd/volumes" Mar 21 05:16:37 crc kubenswrapper[4775]: I0321 05:16:37.040299 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-x5t7w"] Mar 21 05:16:37 crc kubenswrapper[4775]: I0321 05:16:37.052922 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b34c-account-create-update-6kvs4"] Mar 21 05:16:37 crc kubenswrapper[4775]: I0321 05:16:37.067300 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-b34c-account-create-update-6kvs4"] Mar 21 05:16:37 crc kubenswrapper[4775]: I0321 05:16:37.079680 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-x5t7w"] Mar 21 05:16:37 crc kubenswrapper[4775]: I0321 05:16:37.091140 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-6ca3-account-create-update-p8m7j"] Mar 21 05:16:37 crc kubenswrapper[4775]: I0321 05:16:37.100850 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-7a73-account-create-update-jvhc8"] Mar 21 05:16:37 crc kubenswrapper[4775]: I0321 05:16:37.110391 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-rbcrx"] Mar 21 05:16:37 crc kubenswrapper[4775]: I0321 05:16:37.122625 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-7a73-account-create-update-jvhc8"] Mar 21 05:16:37 crc kubenswrapper[4775]: I0321 05:16:37.132381 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-rbcrx"] Mar 21 05:16:37 crc kubenswrapper[4775]: I0321 05:16:37.142307 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-6ca3-account-create-update-p8m7j"] Mar 21 05:16:37 crc kubenswrapper[4775]: I0321 05:16:37.151426 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-9vtvz"] Mar 21 05:16:37 crc kubenswrapper[4775]: I0321 05:16:37.160937 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-9vtvz"] Mar 21 05:16:37 crc kubenswrapper[4775]: I0321 05:16:37.676643 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32831ef2-2e09-453b-9bea-dffe7423fa37" path="/var/lib/kubelet/pods/32831ef2-2e09-453b-9bea-dffe7423fa37/volumes" Mar 21 05:16:37 crc kubenswrapper[4775]: I0321 05:16:37.677432 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74dea94d-fe8b-49e3-9730-fbb641cafaab" path="/var/lib/kubelet/pods/74dea94d-fe8b-49e3-9730-fbb641cafaab/volumes" Mar 21 05:16:37 crc kubenswrapper[4775]: I0321 05:16:37.678145 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="762f4dd9-9a96-4cdb-aa87-e181f5959140" path="/var/lib/kubelet/pods/762f4dd9-9a96-4cdb-aa87-e181f5959140/volumes" Mar 21 05:16:37 crc kubenswrapper[4775]: I0321 05:16:37.678768 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6" path="/var/lib/kubelet/pods/94ac9553-1ce0-4c22-9b2a-e7f38b75a6d6/volumes" Mar 21 05:16:37 crc kubenswrapper[4775]: I0321 05:16:37.679939 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b" path="/var/lib/kubelet/pods/9e291ad2-f2c9-4fd6-8ac8-f1e8e53e6a6b/volumes" Mar 21 05:16:37 crc kubenswrapper[4775]: I0321 05:16:37.680636 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3e27f70-6ea1-4f2e-a42e-e65c8ba76147" path="/var/lib/kubelet/pods/d3e27f70-6ea1-4f2e-a42e-e65c8ba76147/volumes" Mar 21 05:16:39 crc kubenswrapper[4775]: I0321 05:16:39.662386 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:16:39 crc kubenswrapper[4775]: E0321 05:16:39.663644 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:16:44 crc kubenswrapper[4775]: I0321 05:16:44.612728 4775 scope.go:117] "RemoveContainer" containerID="5d0abdc2d7bcbfcb69006c6e61f11899c9c6ba0f99f80bd6ef0c79dc4299f856" Mar 21 05:16:44 crc kubenswrapper[4775]: I0321 05:16:44.649255 4775 scope.go:117] "RemoveContainer" containerID="f3bbbe70d0549a7acec5479ed8da7aa8abe15ef1981c7636975b0a1dd52d35cc" Mar 21 05:16:44 crc kubenswrapper[4775]: I0321 05:16:44.701944 4775 scope.go:117] "RemoveContainer" containerID="3d3602be9c806dde89f844d57f5ed730a681b0e2d28bcdfcf3fd72611d393bf0" Mar 21 05:16:44 crc kubenswrapper[4775]: I0321 05:16:44.743936 4775 scope.go:117] "RemoveContainer" containerID="05fed25f10a4568226dd529e50e8f49b33dd0e98b0199020ea806a0cb77f54ee" Mar 21 05:16:44 crc kubenswrapper[4775]: I0321 05:16:44.792084 4775 scope.go:117] "RemoveContainer" containerID="612cd0e0df303a73de827afa0ae711fb6535100e6b3d945953567e8f21be3662" Mar 21 05:16:44 crc kubenswrapper[4775]: I0321 05:16:44.801541 4775 generic.go:334] "Generic (PLEG): container finished" podID="28040d61-c9ea-4a55-b113-db871dff679c" containerID="4f6b82cb9ea1aa1f200cad1ffb7a707dc8d8464ef48ddbc1ece993160bc193ff" exitCode=0 Mar 21 05:16:44 crc kubenswrapper[4775]: I0321 05:16:44.801637 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr" event={"ID":"28040d61-c9ea-4a55-b113-db871dff679c","Type":"ContainerDied","Data":"4f6b82cb9ea1aa1f200cad1ffb7a707dc8d8464ef48ddbc1ece993160bc193ff"} Mar 21 05:16:44 crc kubenswrapper[4775]: I0321 05:16:44.870085 4775 scope.go:117] "RemoveContainer" containerID="db02c0eaade2846804c5f5491c8f4869f16bd0a0624e9878aa43eba4fd63b5ed" Mar 21 05:16:44 crc kubenswrapper[4775]: I0321 05:16:44.912744 4775 scope.go:117] "RemoveContainer" containerID="96a2bef185388797d3fea57d8ac4fefdb91c53837dbce4eb9abcae899aab9b04" Mar 21 05:16:44 crc kubenswrapper[4775]: I0321 05:16:44.956048 4775 scope.go:117] "RemoveContainer" containerID="7ca2f730d3508ebef97a36404de20e26217deca533b74b90b3415ce6b5463328" Mar 21 05:16:44 crc kubenswrapper[4775]: I0321 05:16:44.990047 4775 scope.go:117] "RemoveContainer" containerID="e4255198ffff3f324a60a06814346bd92cba238dd6b295fce65cd1387f10c2d6" Mar 21 05:16:45 crc kubenswrapper[4775]: I0321 05:16:45.025873 4775 scope.go:117] "RemoveContainer" containerID="2b7a69f7824c3af4bbbbe389045d25783f39861d479383b702bdac257cb7b2e3" Mar 21 05:16:45 crc kubenswrapper[4775]: I0321 05:16:45.049690 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-dl5n2"] Mar 21 05:16:45 crc kubenswrapper[4775]: I0321 05:16:45.061396 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-dl5n2"] Mar 21 05:16:45 crc kubenswrapper[4775]: I0321 05:16:45.064294 4775 scope.go:117] "RemoveContainer" containerID="9076af8667b17ded58ed73ceb5a5784857256d7d90cd1bdaecae68562f5cf370" Mar 21 05:16:45 crc kubenswrapper[4775]: I0321 05:16:45.091197 4775 scope.go:117] "RemoveContainer" containerID="a887c3f98f7514e80429067de51d11f1272a6c7ef6bbe82b5e6e6de5b394be30" Mar 21 05:16:45 crc kubenswrapper[4775]: I0321 05:16:45.117686 4775 scope.go:117] "RemoveContainer" containerID="13179901d0030c432b15b9cdcd3d4bcc8af1fc1231684497164b45d09b84bafe" Mar 21 05:16:45 crc kubenswrapper[4775]: I0321 05:16:45.145035 4775 scope.go:117] "RemoveContainer" containerID="9349131633ffa81b213e3981f82d949940d821489cd39e0d255301316c6d40ec" Mar 21 05:16:45 crc kubenswrapper[4775]: I0321 05:16:45.687500 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47698618-487f-4849-b179-34398850f0e0" path="/var/lib/kubelet/pods/47698618-487f-4849-b179-34398850f0e0/volumes" Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.302501 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr" Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.349782 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28040d61-c9ea-4a55-b113-db871dff679c-inventory\") pod \"28040d61-c9ea-4a55-b113-db871dff679c\" (UID: \"28040d61-c9ea-4a55-b113-db871dff679c\") " Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.350093 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/28040d61-c9ea-4a55-b113-db871dff679c-ssh-key-openstack-edpm-ipam\") pod \"28040d61-c9ea-4a55-b113-db871dff679c\" (UID: \"28040d61-c9ea-4a55-b113-db871dff679c\") " Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.350280 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ps8td\" (UniqueName: \"kubernetes.io/projected/28040d61-c9ea-4a55-b113-db871dff679c-kube-api-access-ps8td\") pod \"28040d61-c9ea-4a55-b113-db871dff679c\" (UID: \"28040d61-c9ea-4a55-b113-db871dff679c\") " Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.356987 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28040d61-c9ea-4a55-b113-db871dff679c-kube-api-access-ps8td" (OuterVolumeSpecName: "kube-api-access-ps8td") pod "28040d61-c9ea-4a55-b113-db871dff679c" (UID: "28040d61-c9ea-4a55-b113-db871dff679c"). InnerVolumeSpecName "kube-api-access-ps8td". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:16:46 crc kubenswrapper[4775]: E0321 05:16:46.384405 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/28040d61-c9ea-4a55-b113-db871dff679c-inventory podName:28040d61-c9ea-4a55-b113-db871dff679c nodeName:}" failed. No retries permitted until 2026-03-21 05:16:46.884379963 +0000 UTC m=+1759.860843587 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/28040d61-c9ea-4a55-b113-db871dff679c-inventory") pod "28040d61-c9ea-4a55-b113-db871dff679c" (UID: "28040d61-c9ea-4a55-b113-db871dff679c") : error deleting /var/lib/kubelet/pods/28040d61-c9ea-4a55-b113-db871dff679c/volume-subpaths: remove /var/lib/kubelet/pods/28040d61-c9ea-4a55-b113-db871dff679c/volume-subpaths: no such file or directory Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.387564 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28040d61-c9ea-4a55-b113-db871dff679c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "28040d61-c9ea-4a55-b113-db871dff679c" (UID: "28040d61-c9ea-4a55-b113-db871dff679c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.452838 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/28040d61-c9ea-4a55-b113-db871dff679c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.452884 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ps8td\" (UniqueName: \"kubernetes.io/projected/28040d61-c9ea-4a55-b113-db871dff679c-kube-api-access-ps8td\") on node \"crc\" DevicePath \"\"" Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.830289 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr" event={"ID":"28040d61-c9ea-4a55-b113-db871dff679c","Type":"ContainerDied","Data":"ef0fcdf856bb63344889100ab23ff6e3c752ecddd8029d7da572bb02105b88a6"} Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.830355 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef0fcdf856bb63344889100ab23ff6e3c752ecddd8029d7da572bb02105b88a6" Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.830385 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-h87nr" Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.917346 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7"] Mar 21 05:16:46 crc kubenswrapper[4775]: E0321 05:16:46.917828 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28040d61-c9ea-4a55-b113-db871dff679c" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.917849 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="28040d61-c9ea-4a55-b113-db871dff679c" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 21 05:16:46 crc kubenswrapper[4775]: E0321 05:16:46.917863 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc" containerName="oc" Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.917870 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc" containerName="oc" Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.918053 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="28040d61-c9ea-4a55-b113-db871dff679c" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.918078 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc" containerName="oc" Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.918819 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7" Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.940064 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7"] Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.967937 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28040d61-c9ea-4a55-b113-db871dff679c-inventory\") pod \"28040d61-c9ea-4a55-b113-db871dff679c\" (UID: \"28040d61-c9ea-4a55-b113-db871dff679c\") " Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.968707 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5efe4255-484c-47d7-800a-4d0dbc5cecd9-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7\" (UID: \"5efe4255-484c-47d7-800a-4d0dbc5cecd9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7" Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.968902 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpgdj\" (UniqueName: \"kubernetes.io/projected/5efe4255-484c-47d7-800a-4d0dbc5cecd9-kube-api-access-qpgdj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7\" (UID: \"5efe4255-484c-47d7-800a-4d0dbc5cecd9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7" Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.969012 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5efe4255-484c-47d7-800a-4d0dbc5cecd9-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7\" (UID: \"5efe4255-484c-47d7-800a-4d0dbc5cecd9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7" Mar 21 05:16:46 crc kubenswrapper[4775]: I0321 05:16:46.976898 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28040d61-c9ea-4a55-b113-db871dff679c-inventory" (OuterVolumeSpecName: "inventory") pod "28040d61-c9ea-4a55-b113-db871dff679c" (UID: "28040d61-c9ea-4a55-b113-db871dff679c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:16:47 crc kubenswrapper[4775]: I0321 05:16:47.071168 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpgdj\" (UniqueName: \"kubernetes.io/projected/5efe4255-484c-47d7-800a-4d0dbc5cecd9-kube-api-access-qpgdj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7\" (UID: \"5efe4255-484c-47d7-800a-4d0dbc5cecd9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7" Mar 21 05:16:47 crc kubenswrapper[4775]: I0321 05:16:47.071314 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5efe4255-484c-47d7-800a-4d0dbc5cecd9-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7\" (UID: \"5efe4255-484c-47d7-800a-4d0dbc5cecd9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7" Mar 21 05:16:47 crc kubenswrapper[4775]: I0321 05:16:47.071457 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5efe4255-484c-47d7-800a-4d0dbc5cecd9-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7\" (UID: \"5efe4255-484c-47d7-800a-4d0dbc5cecd9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7" Mar 21 05:16:47 crc kubenswrapper[4775]: I0321 05:16:47.071573 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28040d61-c9ea-4a55-b113-db871dff679c-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 05:16:47 crc kubenswrapper[4775]: I0321 05:16:47.075487 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5efe4255-484c-47d7-800a-4d0dbc5cecd9-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7\" (UID: \"5efe4255-484c-47d7-800a-4d0dbc5cecd9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7" Mar 21 05:16:47 crc kubenswrapper[4775]: I0321 05:16:47.075487 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5efe4255-484c-47d7-800a-4d0dbc5cecd9-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7\" (UID: \"5efe4255-484c-47d7-800a-4d0dbc5cecd9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7" Mar 21 05:16:47 crc kubenswrapper[4775]: I0321 05:16:47.090593 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpgdj\" (UniqueName: \"kubernetes.io/projected/5efe4255-484c-47d7-800a-4d0dbc5cecd9-kube-api-access-qpgdj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7\" (UID: \"5efe4255-484c-47d7-800a-4d0dbc5cecd9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7" Mar 21 05:16:47 crc kubenswrapper[4775]: I0321 05:16:47.241064 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7" Mar 21 05:16:47 crc kubenswrapper[4775]: I0321 05:16:47.850406 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7"] Mar 21 05:16:47 crc kubenswrapper[4775]: W0321 05:16:47.863659 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5efe4255_484c_47d7_800a_4d0dbc5cecd9.slice/crio-3d0bf76bdd87a405c59bbe5ef7257f5d0e9718a3bd77269f38cb3200a09aee39 WatchSource:0}: Error finding container 3d0bf76bdd87a405c59bbe5ef7257f5d0e9718a3bd77269f38cb3200a09aee39: Status 404 returned error can't find the container with id 3d0bf76bdd87a405c59bbe5ef7257f5d0e9718a3bd77269f38cb3200a09aee39 Mar 21 05:16:48 crc kubenswrapper[4775]: I0321 05:16:48.866965 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7" event={"ID":"5efe4255-484c-47d7-800a-4d0dbc5cecd9","Type":"ContainerStarted","Data":"3d0bf76bdd87a405c59bbe5ef7257f5d0e9718a3bd77269f38cb3200a09aee39"} Mar 21 05:16:49 crc kubenswrapper[4775]: I0321 05:16:49.878488 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7" event={"ID":"5efe4255-484c-47d7-800a-4d0dbc5cecd9","Type":"ContainerStarted","Data":"80480f5e0585d0b76dfc565ed1e8ddada97469b8a016e2301998dae453d6703e"} Mar 21 05:16:49 crc kubenswrapper[4775]: I0321 05:16:49.908164 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7" podStartSLOduration=3.119050994 podStartE2EDuration="3.908113626s" podCreationTimestamp="2026-03-21 05:16:46 +0000 UTC" firstStartedPulling="2026-03-21 05:16:47.879383674 +0000 UTC m=+1760.855847308" lastFinishedPulling="2026-03-21 05:16:48.668446316 +0000 UTC m=+1761.644909940" observedRunningTime="2026-03-21 05:16:49.89648076 +0000 UTC m=+1762.872944404" watchObservedRunningTime="2026-03-21 05:16:49.908113626 +0000 UTC m=+1762.884577250" Mar 21 05:16:51 crc kubenswrapper[4775]: I0321 05:16:51.661649 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:16:51 crc kubenswrapper[4775]: E0321 05:16:51.662516 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:17:02 crc kubenswrapper[4775]: I0321 05:17:02.662469 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:17:02 crc kubenswrapper[4775]: E0321 05:17:02.663872 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:17:15 crc kubenswrapper[4775]: I0321 05:17:15.661884 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:17:15 crc kubenswrapper[4775]: E0321 05:17:15.663048 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:17:27 crc kubenswrapper[4775]: I0321 05:17:27.670014 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:17:27 crc kubenswrapper[4775]: E0321 05:17:27.674039 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:17:34 crc kubenswrapper[4775]: I0321 05:17:34.067264 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-td76j"] Mar 21 05:17:34 crc kubenswrapper[4775]: I0321 05:17:34.080760 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-td76j"] Mar 21 05:17:35 crc kubenswrapper[4775]: I0321 05:17:35.040933 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-jxmtf"] Mar 21 05:17:35 crc kubenswrapper[4775]: I0321 05:17:35.052652 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-jxmtf"] Mar 21 05:17:35 crc kubenswrapper[4775]: I0321 05:17:35.674906 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="716605f1-5111-4e7a-9591-18dfb5da1984" path="/var/lib/kubelet/pods/716605f1-5111-4e7a-9591-18dfb5da1984/volumes" Mar 21 05:17:35 crc kubenswrapper[4775]: I0321 05:17:35.675756 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdf8b79b-6156-4aa5-a769-3a96408745c1" path="/var/lib/kubelet/pods/cdf8b79b-6156-4aa5-a769-3a96408745c1/volumes" Mar 21 05:17:40 crc kubenswrapper[4775]: I0321 05:17:40.662212 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:17:40 crc kubenswrapper[4775]: E0321 05:17:40.663732 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:17:45 crc kubenswrapper[4775]: I0321 05:17:45.425969 4775 scope.go:117] "RemoveContainer" containerID="36bd24e1b5566dbbfb969007cb73ea97d71cecb278d83fa7485e1f5061b3eb1d" Mar 21 05:17:45 crc kubenswrapper[4775]: I0321 05:17:45.469293 4775 scope.go:117] "RemoveContainer" containerID="e72ab47d319ea50893e1c6b3a16586fa25056fa334f81da8dd2632328d94d338" Mar 21 05:17:45 crc kubenswrapper[4775]: I0321 05:17:45.524644 4775 scope.go:117] "RemoveContainer" containerID="43def41a0cf1f318d43db6319e03e84d1a862d118a43d560e7c5e1d51f26dead" Mar 21 05:17:48 crc kubenswrapper[4775]: I0321 05:17:48.040772 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-dh4pj"] Mar 21 05:17:48 crc kubenswrapper[4775]: I0321 05:17:48.060279 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-wvjlb"] Mar 21 05:17:48 crc kubenswrapper[4775]: I0321 05:17:48.071576 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-dh4pj"] Mar 21 05:17:48 crc kubenswrapper[4775]: I0321 05:17:48.080309 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-wvjlb"] Mar 21 05:17:49 crc kubenswrapper[4775]: I0321 05:17:49.673326 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cd9d81f-8c8a-46f9-9943-42dc0b638bef" path="/var/lib/kubelet/pods/5cd9d81f-8c8a-46f9-9943-42dc0b638bef/volumes" Mar 21 05:17:49 crc kubenswrapper[4775]: I0321 05:17:49.674403 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9531f5f9-8f77-4882-b779-4210b1de81de" path="/var/lib/kubelet/pods/9531f5f9-8f77-4882-b779-4210b1de81de/volumes" Mar 21 05:17:50 crc kubenswrapper[4775]: I0321 05:17:50.041287 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-l8qqc"] Mar 21 05:17:50 crc kubenswrapper[4775]: I0321 05:17:50.051810 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-l8qqc"] Mar 21 05:17:51 crc kubenswrapper[4775]: I0321 05:17:51.673875 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37bb7e34-ac47-44f6-b18f-ef4ed78eea6a" path="/var/lib/kubelet/pods/37bb7e34-ac47-44f6-b18f-ef4ed78eea6a/volumes" Mar 21 05:17:53 crc kubenswrapper[4775]: I0321 05:17:53.662312 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:17:53 crc kubenswrapper[4775]: E0321 05:17:53.663577 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:17:55 crc kubenswrapper[4775]: I0321 05:17:55.048007 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-9cmq2"] Mar 21 05:17:55 crc kubenswrapper[4775]: I0321 05:17:55.060181 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-9cmq2"] Mar 21 05:17:55 crc kubenswrapper[4775]: I0321 05:17:55.672114 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1658991f-7a2b-4ce9-a240-a940385e0b8f" path="/var/lib/kubelet/pods/1658991f-7a2b-4ce9-a240-a940385e0b8f/volumes" Mar 21 05:18:00 crc kubenswrapper[4775]: I0321 05:18:00.179318 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567838-kc6hr"] Mar 21 05:18:00 crc kubenswrapper[4775]: I0321 05:18:00.183044 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567838-kc6hr" Mar 21 05:18:00 crc kubenswrapper[4775]: I0321 05:18:00.208442 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:18:00 crc kubenswrapper[4775]: I0321 05:18:00.208815 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:18:00 crc kubenswrapper[4775]: I0321 05:18:00.209159 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:18:00 crc kubenswrapper[4775]: I0321 05:18:00.272882 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567838-kc6hr"] Mar 21 05:18:00 crc kubenswrapper[4775]: I0321 05:18:00.334578 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmn5q\" (UniqueName: \"kubernetes.io/projected/8ed18838-ce8f-499e-a043-886b7e878eb3-kube-api-access-vmn5q\") pod \"auto-csr-approver-29567838-kc6hr\" (UID: \"8ed18838-ce8f-499e-a043-886b7e878eb3\") " pod="openshift-infra/auto-csr-approver-29567838-kc6hr" Mar 21 05:18:00 crc kubenswrapper[4775]: I0321 05:18:00.437478 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmn5q\" (UniqueName: \"kubernetes.io/projected/8ed18838-ce8f-499e-a043-886b7e878eb3-kube-api-access-vmn5q\") pod \"auto-csr-approver-29567838-kc6hr\" (UID: \"8ed18838-ce8f-499e-a043-886b7e878eb3\") " pod="openshift-infra/auto-csr-approver-29567838-kc6hr" Mar 21 05:18:00 crc kubenswrapper[4775]: I0321 05:18:00.461342 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmn5q\" (UniqueName: \"kubernetes.io/projected/8ed18838-ce8f-499e-a043-886b7e878eb3-kube-api-access-vmn5q\") pod \"auto-csr-approver-29567838-kc6hr\" (UID: \"8ed18838-ce8f-499e-a043-886b7e878eb3\") " pod="openshift-infra/auto-csr-approver-29567838-kc6hr" Mar 21 05:18:00 crc kubenswrapper[4775]: I0321 05:18:00.569309 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567838-kc6hr" Mar 21 05:18:01 crc kubenswrapper[4775]: I0321 05:18:01.100706 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567838-kc6hr"] Mar 21 05:18:01 crc kubenswrapper[4775]: I0321 05:18:01.735210 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567838-kc6hr" event={"ID":"8ed18838-ce8f-499e-a043-886b7e878eb3","Type":"ContainerStarted","Data":"91f1137d357579580cb3c8e9d35f7f77751d4d82d0d8668e688013f61047e11a"} Mar 21 05:18:04 crc kubenswrapper[4775]: I0321 05:18:04.772552 4775 generic.go:334] "Generic (PLEG): container finished" podID="8ed18838-ce8f-499e-a043-886b7e878eb3" containerID="f7a7cb2d632bd295102216dc1df2efdc817134357d27ae93440d6f95f32d6b6b" exitCode=0 Mar 21 05:18:04 crc kubenswrapper[4775]: I0321 05:18:04.773017 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567838-kc6hr" event={"ID":"8ed18838-ce8f-499e-a043-886b7e878eb3","Type":"ContainerDied","Data":"f7a7cb2d632bd295102216dc1df2efdc817134357d27ae93440d6f95f32d6b6b"} Mar 21 05:18:05 crc kubenswrapper[4775]: I0321 05:18:05.784338 4775 generic.go:334] "Generic (PLEG): container finished" podID="5efe4255-484c-47d7-800a-4d0dbc5cecd9" containerID="80480f5e0585d0b76dfc565ed1e8ddada97469b8a016e2301998dae453d6703e" exitCode=0 Mar 21 05:18:05 crc kubenswrapper[4775]: I0321 05:18:05.784850 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7" event={"ID":"5efe4255-484c-47d7-800a-4d0dbc5cecd9","Type":"ContainerDied","Data":"80480f5e0585d0b76dfc565ed1e8ddada97469b8a016e2301998dae453d6703e"} Mar 21 05:18:06 crc kubenswrapper[4775]: I0321 05:18:06.183481 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567838-kc6hr" Mar 21 05:18:06 crc kubenswrapper[4775]: I0321 05:18:06.280230 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmn5q\" (UniqueName: \"kubernetes.io/projected/8ed18838-ce8f-499e-a043-886b7e878eb3-kube-api-access-vmn5q\") pod \"8ed18838-ce8f-499e-a043-886b7e878eb3\" (UID: \"8ed18838-ce8f-499e-a043-886b7e878eb3\") " Mar 21 05:18:06 crc kubenswrapper[4775]: I0321 05:18:06.285904 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ed18838-ce8f-499e-a043-886b7e878eb3-kube-api-access-vmn5q" (OuterVolumeSpecName: "kube-api-access-vmn5q") pod "8ed18838-ce8f-499e-a043-886b7e878eb3" (UID: "8ed18838-ce8f-499e-a043-886b7e878eb3"). InnerVolumeSpecName "kube-api-access-vmn5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:18:06 crc kubenswrapper[4775]: I0321 05:18:06.382882 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmn5q\" (UniqueName: \"kubernetes.io/projected/8ed18838-ce8f-499e-a043-886b7e878eb3-kube-api-access-vmn5q\") on node \"crc\" DevicePath \"\"" Mar 21 05:18:06 crc kubenswrapper[4775]: I0321 05:18:06.798705 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567838-kc6hr" Mar 21 05:18:06 crc kubenswrapper[4775]: I0321 05:18:06.798649 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567838-kc6hr" event={"ID":"8ed18838-ce8f-499e-a043-886b7e878eb3","Type":"ContainerDied","Data":"91f1137d357579580cb3c8e9d35f7f77751d4d82d0d8668e688013f61047e11a"} Mar 21 05:18:06 crc kubenswrapper[4775]: I0321 05:18:06.798850 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91f1137d357579580cb3c8e9d35f7f77751d4d82d0d8668e688013f61047e11a" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.274903 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567832-cm8v6"] Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.285781 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567832-cm8v6"] Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.313997 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.442948 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5efe4255-484c-47d7-800a-4d0dbc5cecd9-ssh-key-openstack-edpm-ipam\") pod \"5efe4255-484c-47d7-800a-4d0dbc5cecd9\" (UID: \"5efe4255-484c-47d7-800a-4d0dbc5cecd9\") " Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.443092 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5efe4255-484c-47d7-800a-4d0dbc5cecd9-inventory\") pod \"5efe4255-484c-47d7-800a-4d0dbc5cecd9\" (UID: \"5efe4255-484c-47d7-800a-4d0dbc5cecd9\") " Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.443461 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpgdj\" (UniqueName: \"kubernetes.io/projected/5efe4255-484c-47d7-800a-4d0dbc5cecd9-kube-api-access-qpgdj\") pod \"5efe4255-484c-47d7-800a-4d0dbc5cecd9\" (UID: \"5efe4255-484c-47d7-800a-4d0dbc5cecd9\") " Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.449801 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5efe4255-484c-47d7-800a-4d0dbc5cecd9-kube-api-access-qpgdj" (OuterVolumeSpecName: "kube-api-access-qpgdj") pod "5efe4255-484c-47d7-800a-4d0dbc5cecd9" (UID: "5efe4255-484c-47d7-800a-4d0dbc5cecd9"). InnerVolumeSpecName "kube-api-access-qpgdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.474723 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5efe4255-484c-47d7-800a-4d0dbc5cecd9-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5efe4255-484c-47d7-800a-4d0dbc5cecd9" (UID: "5efe4255-484c-47d7-800a-4d0dbc5cecd9"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.480293 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5efe4255-484c-47d7-800a-4d0dbc5cecd9-inventory" (OuterVolumeSpecName: "inventory") pod "5efe4255-484c-47d7-800a-4d0dbc5cecd9" (UID: "5efe4255-484c-47d7-800a-4d0dbc5cecd9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.547062 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5efe4255-484c-47d7-800a-4d0dbc5cecd9-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.547099 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5efe4255-484c-47d7-800a-4d0dbc5cecd9-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.547147 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpgdj\" (UniqueName: \"kubernetes.io/projected/5efe4255-484c-47d7-800a-4d0dbc5cecd9-kube-api-access-qpgdj\") on node \"crc\" DevicePath \"\"" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.673901 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ffbca22-6533-4ade-8f1f-eeeda82f159f" path="/var/lib/kubelet/pods/0ffbca22-6533-4ade-8f1f-eeeda82f159f/volumes" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.815349 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7" event={"ID":"5efe4255-484c-47d7-800a-4d0dbc5cecd9","Type":"ContainerDied","Data":"3d0bf76bdd87a405c59bbe5ef7257f5d0e9718a3bd77269f38cb3200a09aee39"} Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.815400 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d0bf76bdd87a405c59bbe5ef7257f5d0e9718a3bd77269f38cb3200a09aee39" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.815431 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.924001 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7"] Mar 21 05:18:07 crc kubenswrapper[4775]: E0321 05:18:07.925174 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5efe4255-484c-47d7-800a-4d0dbc5cecd9" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.925195 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5efe4255-484c-47d7-800a-4d0dbc5cecd9" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 21 05:18:07 crc kubenswrapper[4775]: E0321 05:18:07.925217 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ed18838-ce8f-499e-a043-886b7e878eb3" containerName="oc" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.925228 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ed18838-ce8f-499e-a043-886b7e878eb3" containerName="oc" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.925540 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ed18838-ce8f-499e-a043-886b7e878eb3" containerName="oc" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.925563 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5efe4255-484c-47d7-800a-4d0dbc5cecd9" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.926501 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.929008 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.929008 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljtt5" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.929917 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.934573 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 05:18:07 crc kubenswrapper[4775]: I0321 05:18:07.938005 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7"] Mar 21 05:18:08 crc kubenswrapper[4775]: I0321 05:18:08.059333 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/268b27f0-a217-459e-9502-7b522ca6fe2c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mfft7\" (UID: \"268b27f0-a217-459e-9502-7b522ca6fe2c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7" Mar 21 05:18:08 crc kubenswrapper[4775]: I0321 05:18:08.059398 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/268b27f0-a217-459e-9502-7b522ca6fe2c-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mfft7\" (UID: \"268b27f0-a217-459e-9502-7b522ca6fe2c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7" Mar 21 05:18:08 crc kubenswrapper[4775]: I0321 05:18:08.059454 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fskrp\" (UniqueName: \"kubernetes.io/projected/268b27f0-a217-459e-9502-7b522ca6fe2c-kube-api-access-fskrp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mfft7\" (UID: \"268b27f0-a217-459e-9502-7b522ca6fe2c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7" Mar 21 05:18:08 crc kubenswrapper[4775]: I0321 05:18:08.163439 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/268b27f0-a217-459e-9502-7b522ca6fe2c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mfft7\" (UID: \"268b27f0-a217-459e-9502-7b522ca6fe2c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7" Mar 21 05:18:08 crc kubenswrapper[4775]: I0321 05:18:08.163582 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/268b27f0-a217-459e-9502-7b522ca6fe2c-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mfft7\" (UID: \"268b27f0-a217-459e-9502-7b522ca6fe2c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7" Mar 21 05:18:08 crc kubenswrapper[4775]: I0321 05:18:08.163782 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fskrp\" (UniqueName: \"kubernetes.io/projected/268b27f0-a217-459e-9502-7b522ca6fe2c-kube-api-access-fskrp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mfft7\" (UID: \"268b27f0-a217-459e-9502-7b522ca6fe2c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7" Mar 21 05:18:08 crc kubenswrapper[4775]: I0321 05:18:08.169950 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/268b27f0-a217-459e-9502-7b522ca6fe2c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mfft7\" (UID: \"268b27f0-a217-459e-9502-7b522ca6fe2c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7" Mar 21 05:18:08 crc kubenswrapper[4775]: I0321 05:18:08.174881 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/268b27f0-a217-459e-9502-7b522ca6fe2c-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mfft7\" (UID: \"268b27f0-a217-459e-9502-7b522ca6fe2c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7" Mar 21 05:18:08 crc kubenswrapper[4775]: I0321 05:18:08.183550 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fskrp\" (UniqueName: \"kubernetes.io/projected/268b27f0-a217-459e-9502-7b522ca6fe2c-kube-api-access-fskrp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mfft7\" (UID: \"268b27f0-a217-459e-9502-7b522ca6fe2c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7" Mar 21 05:18:08 crc kubenswrapper[4775]: I0321 05:18:08.263849 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7" Mar 21 05:18:08 crc kubenswrapper[4775]: I0321 05:18:08.662414 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:18:08 crc kubenswrapper[4775]: E0321 05:18:08.663264 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:18:09 crc kubenswrapper[4775]: I0321 05:18:09.026713 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7"] Mar 21 05:18:09 crc kubenswrapper[4775]: I0321 05:18:09.839948 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7" event={"ID":"268b27f0-a217-459e-9502-7b522ca6fe2c","Type":"ContainerStarted","Data":"220eaeb65dcf0a6aab2fe24c811cc732c89d575f1943b110ba154279a1328887"} Mar 21 05:18:10 crc kubenswrapper[4775]: I0321 05:18:10.854178 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7" event={"ID":"268b27f0-a217-459e-9502-7b522ca6fe2c","Type":"ContainerStarted","Data":"74f1c56742415a3035f0eba080f9203972e8b87038befa94adb04c3bc1be5ffb"} Mar 21 05:18:10 crc kubenswrapper[4775]: I0321 05:18:10.881424 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7" podStartSLOduration=3.410989358 podStartE2EDuration="3.881399908s" podCreationTimestamp="2026-03-21 05:18:07 +0000 UTC" firstStartedPulling="2026-03-21 05:18:09.036517779 +0000 UTC m=+1842.012981393" lastFinishedPulling="2026-03-21 05:18:09.506928319 +0000 UTC m=+1842.483391943" observedRunningTime="2026-03-21 05:18:10.874640018 +0000 UTC m=+1843.851103642" watchObservedRunningTime="2026-03-21 05:18:10.881399908 +0000 UTC m=+1843.857863532" Mar 21 05:18:14 crc kubenswrapper[4775]: I0321 05:18:14.919175 4775 generic.go:334] "Generic (PLEG): container finished" podID="268b27f0-a217-459e-9502-7b522ca6fe2c" containerID="74f1c56742415a3035f0eba080f9203972e8b87038befa94adb04c3bc1be5ffb" exitCode=0 Mar 21 05:18:14 crc kubenswrapper[4775]: I0321 05:18:14.919279 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7" event={"ID":"268b27f0-a217-459e-9502-7b522ca6fe2c","Type":"ContainerDied","Data":"74f1c56742415a3035f0eba080f9203972e8b87038befa94adb04c3bc1be5ffb"} Mar 21 05:18:16 crc kubenswrapper[4775]: I0321 05:18:16.810286 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7" Mar 21 05:18:16 crc kubenswrapper[4775]: I0321 05:18:16.996545 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fskrp\" (UniqueName: \"kubernetes.io/projected/268b27f0-a217-459e-9502-7b522ca6fe2c-kube-api-access-fskrp\") pod \"268b27f0-a217-459e-9502-7b522ca6fe2c\" (UID: \"268b27f0-a217-459e-9502-7b522ca6fe2c\") " Mar 21 05:18:16 crc kubenswrapper[4775]: I0321 05:18:16.996970 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/268b27f0-a217-459e-9502-7b522ca6fe2c-inventory\") pod \"268b27f0-a217-459e-9502-7b522ca6fe2c\" (UID: \"268b27f0-a217-459e-9502-7b522ca6fe2c\") " Mar 21 05:18:16 crc kubenswrapper[4775]: I0321 05:18:16.997140 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/268b27f0-a217-459e-9502-7b522ca6fe2c-ssh-key-openstack-edpm-ipam\") pod \"268b27f0-a217-459e-9502-7b522ca6fe2c\" (UID: \"268b27f0-a217-459e-9502-7b522ca6fe2c\") " Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.015628 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/268b27f0-a217-459e-9502-7b522ca6fe2c-kube-api-access-fskrp" (OuterVolumeSpecName: "kube-api-access-fskrp") pod "268b27f0-a217-459e-9502-7b522ca6fe2c" (UID: "268b27f0-a217-459e-9502-7b522ca6fe2c"). InnerVolumeSpecName "kube-api-access-fskrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.046055 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2"] Mar 21 05:18:17 crc kubenswrapper[4775]: E0321 05:18:17.046608 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="268b27f0-a217-459e-9502-7b522ca6fe2c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.046628 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="268b27f0-a217-459e-9502-7b522ca6fe2c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.046826 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="268b27f0-a217-459e-9502-7b522ca6fe2c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.047646 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.047866 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/268b27f0-a217-459e-9502-7b522ca6fe2c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "268b27f0-a217-459e-9502-7b522ca6fe2c" (UID: "268b27f0-a217-459e-9502-7b522ca6fe2c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.051377 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/268b27f0-a217-459e-9502-7b522ca6fe2c-inventory" (OuterVolumeSpecName: "inventory") pod "268b27f0-a217-459e-9502-7b522ca6fe2c" (UID: "268b27f0-a217-459e-9502-7b522ca6fe2c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.081418 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2"] Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.100076 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fskrp\" (UniqueName: \"kubernetes.io/projected/268b27f0-a217-459e-9502-7b522ca6fe2c-kube-api-access-fskrp\") on node \"crc\" DevicePath \"\"" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.100135 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/268b27f0-a217-459e-9502-7b522ca6fe2c-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.100149 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/268b27f0-a217-459e-9502-7b522ca6fe2c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.155467 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7" event={"ID":"268b27f0-a217-459e-9502-7b522ca6fe2c","Type":"ContainerDied","Data":"220eaeb65dcf0a6aab2fe24c811cc732c89d575f1943b110ba154279a1328887"} Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.155537 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="220eaeb65dcf0a6aab2fe24c811cc732c89d575f1943b110ba154279a1328887" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.155552 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mfft7" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.202794 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbd64e65-be8d-42e7-a686-d5454932156d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kfrz2\" (UID: \"dbd64e65-be8d-42e7-a686-d5454932156d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.203096 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dbd64e65-be8d-42e7-a686-d5454932156d-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kfrz2\" (UID: \"dbd64e65-be8d-42e7-a686-d5454932156d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.203336 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znk7t\" (UniqueName: \"kubernetes.io/projected/dbd64e65-be8d-42e7-a686-d5454932156d-kube-api-access-znk7t\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kfrz2\" (UID: \"dbd64e65-be8d-42e7-a686-d5454932156d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.306199 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znk7t\" (UniqueName: \"kubernetes.io/projected/dbd64e65-be8d-42e7-a686-d5454932156d-kube-api-access-znk7t\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kfrz2\" (UID: \"dbd64e65-be8d-42e7-a686-d5454932156d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.306285 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbd64e65-be8d-42e7-a686-d5454932156d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kfrz2\" (UID: \"dbd64e65-be8d-42e7-a686-d5454932156d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.306452 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dbd64e65-be8d-42e7-a686-d5454932156d-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kfrz2\" (UID: \"dbd64e65-be8d-42e7-a686-d5454932156d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.311763 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbd64e65-be8d-42e7-a686-d5454932156d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kfrz2\" (UID: \"dbd64e65-be8d-42e7-a686-d5454932156d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.311799 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dbd64e65-be8d-42e7-a686-d5454932156d-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kfrz2\" (UID: \"dbd64e65-be8d-42e7-a686-d5454932156d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.325502 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znk7t\" (UniqueName: \"kubernetes.io/projected/dbd64e65-be8d-42e7-a686-d5454932156d-kube-api-access-znk7t\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kfrz2\" (UID: \"dbd64e65-be8d-42e7-a686-d5454932156d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.463493 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2" Mar 21 05:18:17 crc kubenswrapper[4775]: I0321 05:18:17.996558 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2"] Mar 21 05:18:18 crc kubenswrapper[4775]: I0321 05:18:18.169381 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2" event={"ID":"dbd64e65-be8d-42e7-a686-d5454932156d","Type":"ContainerStarted","Data":"051257f240552b19775c8d22bcf0e38507e6f1d72aceca09aa4be651f10da4e9"} Mar 21 05:18:19 crc kubenswrapper[4775]: I0321 05:18:19.184402 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2" event={"ID":"dbd64e65-be8d-42e7-a686-d5454932156d","Type":"ContainerStarted","Data":"799cfa377f0b8d825e50cec86644f9bddd2d5dcd591a4f502902d1ae672bddca"} Mar 21 05:18:19 crc kubenswrapper[4775]: I0321 05:18:19.211010 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2" podStartSLOduration=1.675271773 podStartE2EDuration="2.210989029s" podCreationTimestamp="2026-03-21 05:18:17 +0000 UTC" firstStartedPulling="2026-03-21 05:18:18.018478044 +0000 UTC m=+1850.994941668" lastFinishedPulling="2026-03-21 05:18:18.5541953 +0000 UTC m=+1851.530658924" observedRunningTime="2026-03-21 05:18:19.209338413 +0000 UTC m=+1852.185802047" watchObservedRunningTime="2026-03-21 05:18:19.210989029 +0000 UTC m=+1852.187452653" Mar 21 05:18:23 crc kubenswrapper[4775]: I0321 05:18:23.662895 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:18:23 crc kubenswrapper[4775]: E0321 05:18:23.665017 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:18:30 crc kubenswrapper[4775]: I0321 05:18:30.055134 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-45fb-account-create-update-glct2"] Mar 21 05:18:30 crc kubenswrapper[4775]: I0321 05:18:30.068026 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-a38c-account-create-update-x226p"] Mar 21 05:18:30 crc kubenswrapper[4775]: I0321 05:18:30.083963 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db3c-account-create-update-k29zp"] Mar 21 05:18:30 crc kubenswrapper[4775]: I0321 05:18:30.094705 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-tmm94"] Mar 21 05:18:30 crc kubenswrapper[4775]: I0321 05:18:30.103582 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-zfz4w"] Mar 21 05:18:30 crc kubenswrapper[4775]: I0321 05:18:30.112035 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-45fb-account-create-update-glct2"] Mar 21 05:18:30 crc kubenswrapper[4775]: I0321 05:18:30.123404 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db3c-account-create-update-k29zp"] Mar 21 05:18:30 crc kubenswrapper[4775]: I0321 05:18:30.136743 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-wxmkg"] Mar 21 05:18:30 crc kubenswrapper[4775]: I0321 05:18:30.144493 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-zfz4w"] Mar 21 05:18:30 crc kubenswrapper[4775]: I0321 05:18:30.151930 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-tmm94"] Mar 21 05:18:30 crc kubenswrapper[4775]: I0321 05:18:30.160754 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-a38c-account-create-update-x226p"] Mar 21 05:18:30 crc kubenswrapper[4775]: I0321 05:18:30.169337 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-wxmkg"] Mar 21 05:18:31 crc kubenswrapper[4775]: I0321 05:18:31.675017 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="028d46f7-6f14-40f6-a6cf-77aeec59a99e" path="/var/lib/kubelet/pods/028d46f7-6f14-40f6-a6cf-77aeec59a99e/volumes" Mar 21 05:18:31 crc kubenswrapper[4775]: I0321 05:18:31.677806 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22a5e145-5b67-4d02-9104-ade3e48888db" path="/var/lib/kubelet/pods/22a5e145-5b67-4d02-9104-ade3e48888db/volumes" Mar 21 05:18:31 crc kubenswrapper[4775]: I0321 05:18:31.678482 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="330e8d60-7af3-42ed-a0d7-234264037d09" path="/var/lib/kubelet/pods/330e8d60-7af3-42ed-a0d7-234264037d09/volumes" Mar 21 05:18:31 crc kubenswrapper[4775]: I0321 05:18:31.679211 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87e82da2-fdd4-4ead-8a9b-022e55a03690" path="/var/lib/kubelet/pods/87e82da2-fdd4-4ead-8a9b-022e55a03690/volumes" Mar 21 05:18:31 crc kubenswrapper[4775]: I0321 05:18:31.680490 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b65ed673-344a-4064-ae90-c9d20964c648" path="/var/lib/kubelet/pods/b65ed673-344a-4064-ae90-c9d20964c648/volumes" Mar 21 05:18:31 crc kubenswrapper[4775]: I0321 05:18:31.681014 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc238c51-34cc-4cd6-bdf2-00e3747a9d67" path="/var/lib/kubelet/pods/bc238c51-34cc-4cd6-bdf2-00e3747a9d67/volumes" Mar 21 05:18:38 crc kubenswrapper[4775]: I0321 05:18:38.662765 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:18:38 crc kubenswrapper[4775]: E0321 05:18:38.664165 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:18:45 crc kubenswrapper[4775]: I0321 05:18:45.657621 4775 scope.go:117] "RemoveContainer" containerID="fdab6584b0a9bddc76fc370c865954a1758e24e006007d8b2953684194b46540" Mar 21 05:18:45 crc kubenswrapper[4775]: I0321 05:18:45.722295 4775 scope.go:117] "RemoveContainer" containerID="eb428452fe127829b5fdd3edea1e21f849f0fb951673f81da2334a3d27807ded" Mar 21 05:18:45 crc kubenswrapper[4775]: I0321 05:18:45.761203 4775 scope.go:117] "RemoveContainer" containerID="1a60b462497324b6ca2e30820dbc175100c3b70b7e0a4ec6225ceb18cce50eaa" Mar 21 05:18:45 crc kubenswrapper[4775]: I0321 05:18:45.821971 4775 scope.go:117] "RemoveContainer" containerID="647e5b30ad956a8a9a631808f860ad303cf32d4aa053654a5cdcf1129262af27" Mar 21 05:18:45 crc kubenswrapper[4775]: I0321 05:18:45.852876 4775 scope.go:117] "RemoveContainer" containerID="f761be7525081d117b389918983117196fdba3258fd2ecd1f48a78a479770d1c" Mar 21 05:18:45 crc kubenswrapper[4775]: I0321 05:18:45.914564 4775 scope.go:117] "RemoveContainer" containerID="796b087daa0094c51c318e24d71b74b2beb873149bcfc37bff7fc5e4bd3c8ed7" Mar 21 05:18:45 crc kubenswrapper[4775]: I0321 05:18:45.968560 4775 scope.go:117] "RemoveContainer" containerID="959f27ca1f8883c27a7d20345a9d58bad88607efdd52bc8596a92a7571fa819e" Mar 21 05:18:45 crc kubenswrapper[4775]: I0321 05:18:45.995104 4775 scope.go:117] "RemoveContainer" containerID="84b76b926432d603dd6d10be15e6cfc9e20ac65787b9fb2a30c3bdbd38a96e09" Mar 21 05:18:46 crc kubenswrapper[4775]: I0321 05:18:46.048385 4775 scope.go:117] "RemoveContainer" containerID="f19a3eb592d811463bd26fefd306c16b1ec6187913716e5585ce5b02489e2c0c" Mar 21 05:18:46 crc kubenswrapper[4775]: I0321 05:18:46.102796 4775 scope.go:117] "RemoveContainer" containerID="0ff1159b21db4525f027b0be154a99974ea3e42eac6bac07387f589a08d088ce" Mar 21 05:18:46 crc kubenswrapper[4775]: I0321 05:18:46.131304 4775 scope.go:117] "RemoveContainer" containerID="1fd0a588f9d34ae16dba50e92cfc4f254b48cad183d5ee54a522ba1c5c91886c" Mar 21 05:18:51 crc kubenswrapper[4775]: I0321 05:18:51.663180 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:18:51 crc kubenswrapper[4775]: E0321 05:18:51.664532 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:18:54 crc kubenswrapper[4775]: I0321 05:18:54.592343 4775 generic.go:334] "Generic (PLEG): container finished" podID="dbd64e65-be8d-42e7-a686-d5454932156d" containerID="799cfa377f0b8d825e50cec86644f9bddd2d5dcd591a4f502902d1ae672bddca" exitCode=0 Mar 21 05:18:54 crc kubenswrapper[4775]: I0321 05:18:54.592428 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2" event={"ID":"dbd64e65-be8d-42e7-a686-d5454932156d","Type":"ContainerDied","Data":"799cfa377f0b8d825e50cec86644f9bddd2d5dcd591a4f502902d1ae672bddca"} Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.059060 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gk78h"] Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.070329 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gk78h"] Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.088701 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.143414 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbd64e65-be8d-42e7-a686-d5454932156d-inventory\") pod \"dbd64e65-be8d-42e7-a686-d5454932156d\" (UID: \"dbd64e65-be8d-42e7-a686-d5454932156d\") " Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.143558 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dbd64e65-be8d-42e7-a686-d5454932156d-ssh-key-openstack-edpm-ipam\") pod \"dbd64e65-be8d-42e7-a686-d5454932156d\" (UID: \"dbd64e65-be8d-42e7-a686-d5454932156d\") " Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.143773 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znk7t\" (UniqueName: \"kubernetes.io/projected/dbd64e65-be8d-42e7-a686-d5454932156d-kube-api-access-znk7t\") pod \"dbd64e65-be8d-42e7-a686-d5454932156d\" (UID: \"dbd64e65-be8d-42e7-a686-d5454932156d\") " Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.152575 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbd64e65-be8d-42e7-a686-d5454932156d-kube-api-access-znk7t" (OuterVolumeSpecName: "kube-api-access-znk7t") pod "dbd64e65-be8d-42e7-a686-d5454932156d" (UID: "dbd64e65-be8d-42e7-a686-d5454932156d"). InnerVolumeSpecName "kube-api-access-znk7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.180053 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbd64e65-be8d-42e7-a686-d5454932156d-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "dbd64e65-be8d-42e7-a686-d5454932156d" (UID: "dbd64e65-be8d-42e7-a686-d5454932156d"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.182811 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbd64e65-be8d-42e7-a686-d5454932156d-inventory" (OuterVolumeSpecName: "inventory") pod "dbd64e65-be8d-42e7-a686-d5454932156d" (UID: "dbd64e65-be8d-42e7-a686-d5454932156d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.246353 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dbd64e65-be8d-42e7-a686-d5454932156d-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.246426 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znk7t\" (UniqueName: \"kubernetes.io/projected/dbd64e65-be8d-42e7-a686-d5454932156d-kube-api-access-znk7t\") on node \"crc\" DevicePath \"\"" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.246442 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbd64e65-be8d-42e7-a686-d5454932156d-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.616166 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2" event={"ID":"dbd64e65-be8d-42e7-a686-d5454932156d","Type":"ContainerDied","Data":"051257f240552b19775c8d22bcf0e38507e6f1d72aceca09aa4be651f10da4e9"} Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.616224 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="051257f240552b19775c8d22bcf0e38507e6f1d72aceca09aa4be651f10da4e9" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.616250 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kfrz2" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.722787 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn"] Mar 21 05:18:56 crc kubenswrapper[4775]: E0321 05:18:56.723378 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbd64e65-be8d-42e7-a686-d5454932156d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.723401 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbd64e65-be8d-42e7-a686-d5454932156d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.723611 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbd64e65-be8d-42e7-a686-d5454932156d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.724489 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.731736 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljtt5" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.732139 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.733403 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.735754 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.743763 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn"] Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.756360 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx9km\" (UniqueName: \"kubernetes.io/projected/8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd-kube-api-access-dx9km\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn\" (UID: \"8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.757052 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn\" (UID: \"8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.757136 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn\" (UID: \"8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.860186 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn\" (UID: \"8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.860266 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn\" (UID: \"8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.860339 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx9km\" (UniqueName: \"kubernetes.io/projected/8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd-kube-api-access-dx9km\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn\" (UID: \"8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.867686 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn\" (UID: \"8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.867693 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn\" (UID: \"8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn" Mar 21 05:18:56 crc kubenswrapper[4775]: I0321 05:18:56.893941 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx9km\" (UniqueName: \"kubernetes.io/projected/8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd-kube-api-access-dx9km\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn\" (UID: \"8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn" Mar 21 05:18:57 crc kubenswrapper[4775]: I0321 05:18:57.045167 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn" Mar 21 05:18:57 crc kubenswrapper[4775]: I0321 05:18:57.591026 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn"] Mar 21 05:18:57 crc kubenswrapper[4775]: I0321 05:18:57.628763 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn" event={"ID":"8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd","Type":"ContainerStarted","Data":"8ae2ec9e0825e6dd97cd32fd99b532a4daf6bd5ca6135d8ff0e1f5859b77b50f"} Mar 21 05:18:57 crc kubenswrapper[4775]: I0321 05:18:57.673610 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="824625b1-30cc-42c5-ad83-5854770c2f46" path="/var/lib/kubelet/pods/824625b1-30cc-42c5-ad83-5854770c2f46/volumes" Mar 21 05:18:58 crc kubenswrapper[4775]: I0321 05:18:58.641512 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn" event={"ID":"8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd","Type":"ContainerStarted","Data":"0db68889aac768d03b3c79bf8ab7779d509c671b3324f27b94397e0842b27b48"} Mar 21 05:18:58 crc kubenswrapper[4775]: I0321 05:18:58.671674 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn" podStartSLOduration=2.072325403 podStartE2EDuration="2.671638637s" podCreationTimestamp="2026-03-21 05:18:56 +0000 UTC" firstStartedPulling="2026-03-21 05:18:57.605439451 +0000 UTC m=+1890.581903075" lastFinishedPulling="2026-03-21 05:18:58.204752695 +0000 UTC m=+1891.181216309" observedRunningTime="2026-03-21 05:18:58.659873066 +0000 UTC m=+1891.636336720" watchObservedRunningTime="2026-03-21 05:18:58.671638637 +0000 UTC m=+1891.648102261" Mar 21 05:19:06 crc kubenswrapper[4775]: I0321 05:19:06.661507 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:19:07 crc kubenswrapper[4775]: I0321 05:19:07.742202 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerStarted","Data":"820b5e7971de8f534813cc1fdb277aabd37bf9063cedc671ef2c92d0328150cb"} Mar 21 05:19:18 crc kubenswrapper[4775]: I0321 05:19:18.036699 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-pckp8"] Mar 21 05:19:18 crc kubenswrapper[4775]: I0321 05:19:18.071171 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-pckp8"] Mar 21 05:19:19 crc kubenswrapper[4775]: I0321 05:19:19.673718 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a407dd6-593f-4806-884b-8e031639d25d" path="/var/lib/kubelet/pods/5a407dd6-593f-4806-884b-8e031639d25d/volumes" Mar 21 05:19:21 crc kubenswrapper[4775]: I0321 05:19:21.050209 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vd764"] Mar 21 05:19:21 crc kubenswrapper[4775]: I0321 05:19:21.059773 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vd764"] Mar 21 05:19:21 crc kubenswrapper[4775]: I0321 05:19:21.675928 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb1dd6e9-8801-464e-952c-d345498b132a" path="/var/lib/kubelet/pods/fb1dd6e9-8801-464e-952c-d345498b132a/volumes" Mar 21 05:19:46 crc kubenswrapper[4775]: I0321 05:19:46.394575 4775 scope.go:117] "RemoveContainer" containerID="650b9597595a69528b159834c598c6d1c2eb6858daf6ed558c1a6a03d2d450dd" Mar 21 05:19:46 crc kubenswrapper[4775]: I0321 05:19:46.442836 4775 scope.go:117] "RemoveContainer" containerID="166965bd5533ce39880e58f576322a3fbe5ebbddcee7067e02f0a62bbadf6013" Mar 21 05:19:46 crc kubenswrapper[4775]: I0321 05:19:46.569495 4775 scope.go:117] "RemoveContainer" containerID="20d707fa7f70310d9cb80b150c04d331c2047a421ec95e9ba298af9cad00d19d" Mar 21 05:19:47 crc kubenswrapper[4775]: I0321 05:19:47.227868 4775 generic.go:334] "Generic (PLEG): container finished" podID="8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd" containerID="0db68889aac768d03b3c79bf8ab7779d509c671b3324f27b94397e0842b27b48" exitCode=0 Mar 21 05:19:47 crc kubenswrapper[4775]: I0321 05:19:47.227985 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn" event={"ID":"8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd","Type":"ContainerDied","Data":"0db68889aac768d03b3c79bf8ab7779d509c671b3324f27b94397e0842b27b48"} Mar 21 05:19:48 crc kubenswrapper[4775]: I0321 05:19:48.740995 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn" Mar 21 05:19:48 crc kubenswrapper[4775]: I0321 05:19:48.830300 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd-ssh-key-openstack-edpm-ipam\") pod \"8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd\" (UID: \"8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd\") " Mar 21 05:19:48 crc kubenswrapper[4775]: I0321 05:19:48.830447 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dx9km\" (UniqueName: \"kubernetes.io/projected/8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd-kube-api-access-dx9km\") pod \"8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd\" (UID: \"8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd\") " Mar 21 05:19:48 crc kubenswrapper[4775]: I0321 05:19:48.830498 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd-inventory\") pod \"8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd\" (UID: \"8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd\") " Mar 21 05:19:48 crc kubenswrapper[4775]: I0321 05:19:48.838725 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd-kube-api-access-dx9km" (OuterVolumeSpecName: "kube-api-access-dx9km") pod "8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd" (UID: "8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd"). InnerVolumeSpecName "kube-api-access-dx9km". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:19:48 crc kubenswrapper[4775]: I0321 05:19:48.863967 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd" (UID: "8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:19:48 crc kubenswrapper[4775]: I0321 05:19:48.864394 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd-inventory" (OuterVolumeSpecName: "inventory") pod "8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd" (UID: "8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:19:48 crc kubenswrapper[4775]: I0321 05:19:48.933865 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 05:19:48 crc kubenswrapper[4775]: I0321 05:19:48.933908 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dx9km\" (UniqueName: \"kubernetes.io/projected/8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd-kube-api-access-dx9km\") on node \"crc\" DevicePath \"\"" Mar 21 05:19:48 crc kubenswrapper[4775]: I0321 05:19:48.933923 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.251872 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn" event={"ID":"8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd","Type":"ContainerDied","Data":"8ae2ec9e0825e6dd97cd32fd99b532a4daf6bd5ca6135d8ff0e1f5859b77b50f"} Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.251933 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ae2ec9e0825e6dd97cd32fd99b532a4daf6bd5ca6135d8ff0e1f5859b77b50f" Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.251977 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn" Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.355105 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-swg9k"] Mar 21 05:19:49 crc kubenswrapper[4775]: E0321 05:19:49.355660 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.355682 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.355976 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.356917 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-swg9k" Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.361691 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.361723 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.361701 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljtt5" Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.361910 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.368470 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-swg9k"] Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.447065 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzxn7\" (UniqueName: \"kubernetes.io/projected/5f45d376-4f59-4584-a545-16d4ff066232-kube-api-access-mzxn7\") pod \"ssh-known-hosts-edpm-deployment-swg9k\" (UID: \"5f45d376-4f59-4584-a545-16d4ff066232\") " pod="openstack/ssh-known-hosts-edpm-deployment-swg9k" Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.447317 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5f45d376-4f59-4584-a545-16d4ff066232-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-swg9k\" (UID: \"5f45d376-4f59-4584-a545-16d4ff066232\") " pod="openstack/ssh-known-hosts-edpm-deployment-swg9k" Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.447381 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5f45d376-4f59-4584-a545-16d4ff066232-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-swg9k\" (UID: \"5f45d376-4f59-4584-a545-16d4ff066232\") " pod="openstack/ssh-known-hosts-edpm-deployment-swg9k" Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.549707 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzxn7\" (UniqueName: \"kubernetes.io/projected/5f45d376-4f59-4584-a545-16d4ff066232-kube-api-access-mzxn7\") pod \"ssh-known-hosts-edpm-deployment-swg9k\" (UID: \"5f45d376-4f59-4584-a545-16d4ff066232\") " pod="openstack/ssh-known-hosts-edpm-deployment-swg9k" Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.550306 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5f45d376-4f59-4584-a545-16d4ff066232-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-swg9k\" (UID: \"5f45d376-4f59-4584-a545-16d4ff066232\") " pod="openstack/ssh-known-hosts-edpm-deployment-swg9k" Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.550358 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5f45d376-4f59-4584-a545-16d4ff066232-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-swg9k\" (UID: \"5f45d376-4f59-4584-a545-16d4ff066232\") " pod="openstack/ssh-known-hosts-edpm-deployment-swg9k" Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.557983 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5f45d376-4f59-4584-a545-16d4ff066232-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-swg9k\" (UID: \"5f45d376-4f59-4584-a545-16d4ff066232\") " pod="openstack/ssh-known-hosts-edpm-deployment-swg9k" Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.557981 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5f45d376-4f59-4584-a545-16d4ff066232-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-swg9k\" (UID: \"5f45d376-4f59-4584-a545-16d4ff066232\") " pod="openstack/ssh-known-hosts-edpm-deployment-swg9k" Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.571602 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzxn7\" (UniqueName: \"kubernetes.io/projected/5f45d376-4f59-4584-a545-16d4ff066232-kube-api-access-mzxn7\") pod \"ssh-known-hosts-edpm-deployment-swg9k\" (UID: \"5f45d376-4f59-4584-a545-16d4ff066232\") " pod="openstack/ssh-known-hosts-edpm-deployment-swg9k" Mar 21 05:19:49 crc kubenswrapper[4775]: I0321 05:19:49.692434 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-swg9k" Mar 21 05:19:50 crc kubenswrapper[4775]: I0321 05:19:50.228722 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-swg9k"] Mar 21 05:19:50 crc kubenswrapper[4775]: I0321 05:19:50.244286 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 05:19:50 crc kubenswrapper[4775]: I0321 05:19:50.263430 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-swg9k" event={"ID":"5f45d376-4f59-4584-a545-16d4ff066232","Type":"ContainerStarted","Data":"c8515e9316e1445df316546fb379a9a67b9033ea56ac37c614c85f82c6e2e36b"} Mar 21 05:19:51 crc kubenswrapper[4775]: I0321 05:19:51.288202 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-swg9k" event={"ID":"5f45d376-4f59-4584-a545-16d4ff066232","Type":"ContainerStarted","Data":"cd682044f1ea6f41eb1b64cb63d1eba5cf3b6662c7cf831d0a19d6298440a380"} Mar 21 05:19:58 crc kubenswrapper[4775]: I0321 05:19:58.380254 4775 generic.go:334] "Generic (PLEG): container finished" podID="5f45d376-4f59-4584-a545-16d4ff066232" containerID="cd682044f1ea6f41eb1b64cb63d1eba5cf3b6662c7cf831d0a19d6298440a380" exitCode=0 Mar 21 05:19:58 crc kubenswrapper[4775]: I0321 05:19:58.380334 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-swg9k" event={"ID":"5f45d376-4f59-4584-a545-16d4ff066232","Type":"ContainerDied","Data":"cd682044f1ea6f41eb1b64cb63d1eba5cf3b6662c7cf831d0a19d6298440a380"} Mar 21 05:19:59 crc kubenswrapper[4775]: I0321 05:19:59.885163 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-swg9k" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.000110 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5f45d376-4f59-4584-a545-16d4ff066232-inventory-0\") pod \"5f45d376-4f59-4584-a545-16d4ff066232\" (UID: \"5f45d376-4f59-4584-a545-16d4ff066232\") " Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.000363 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzxn7\" (UniqueName: \"kubernetes.io/projected/5f45d376-4f59-4584-a545-16d4ff066232-kube-api-access-mzxn7\") pod \"5f45d376-4f59-4584-a545-16d4ff066232\" (UID: \"5f45d376-4f59-4584-a545-16d4ff066232\") " Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.000536 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5f45d376-4f59-4584-a545-16d4ff066232-ssh-key-openstack-edpm-ipam\") pod \"5f45d376-4f59-4584-a545-16d4ff066232\" (UID: \"5f45d376-4f59-4584-a545-16d4ff066232\") " Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.008250 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f45d376-4f59-4584-a545-16d4ff066232-kube-api-access-mzxn7" (OuterVolumeSpecName: "kube-api-access-mzxn7") pod "5f45d376-4f59-4584-a545-16d4ff066232" (UID: "5f45d376-4f59-4584-a545-16d4ff066232"). InnerVolumeSpecName "kube-api-access-mzxn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.032412 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f45d376-4f59-4584-a545-16d4ff066232-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "5f45d376-4f59-4584-a545-16d4ff066232" (UID: "5f45d376-4f59-4584-a545-16d4ff066232"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.036632 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f45d376-4f59-4584-a545-16d4ff066232-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5f45d376-4f59-4584-a545-16d4ff066232" (UID: "5f45d376-4f59-4584-a545-16d4ff066232"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.103389 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5f45d376-4f59-4584-a545-16d4ff066232-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.103529 4775 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5f45d376-4f59-4584-a545-16d4ff066232-inventory-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.103544 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzxn7\" (UniqueName: \"kubernetes.io/projected/5f45d376-4f59-4584-a545-16d4ff066232-kube-api-access-mzxn7\") on node \"crc\" DevicePath \"\"" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.165386 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567840-lm4bc"] Mar 21 05:20:00 crc kubenswrapper[4775]: E0321 05:20:00.166101 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f45d376-4f59-4584-a545-16d4ff066232" containerName="ssh-known-hosts-edpm-deployment" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.166141 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f45d376-4f59-4584-a545-16d4ff066232" containerName="ssh-known-hosts-edpm-deployment" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.166439 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f45d376-4f59-4584-a545-16d4ff066232" containerName="ssh-known-hosts-edpm-deployment" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.167891 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567840-lm4bc" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.175048 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.175075 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.175183 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.181624 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567840-lm4bc"] Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.307979 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q2gp\" (UniqueName: \"kubernetes.io/projected/90536c4e-c461-4d2c-95dd-f08660cb2e69-kube-api-access-8q2gp\") pod \"auto-csr-approver-29567840-lm4bc\" (UID: \"90536c4e-c461-4d2c-95dd-f08660cb2e69\") " pod="openshift-infra/auto-csr-approver-29567840-lm4bc" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.403931 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-swg9k" event={"ID":"5f45d376-4f59-4584-a545-16d4ff066232","Type":"ContainerDied","Data":"c8515e9316e1445df316546fb379a9a67b9033ea56ac37c614c85f82c6e2e36b"} Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.403985 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8515e9316e1445df316546fb379a9a67b9033ea56ac37c614c85f82c6e2e36b" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.404055 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-swg9k" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.409921 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q2gp\" (UniqueName: \"kubernetes.io/projected/90536c4e-c461-4d2c-95dd-f08660cb2e69-kube-api-access-8q2gp\") pod \"auto-csr-approver-29567840-lm4bc\" (UID: \"90536c4e-c461-4d2c-95dd-f08660cb2e69\") " pod="openshift-infra/auto-csr-approver-29567840-lm4bc" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.446580 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q2gp\" (UniqueName: \"kubernetes.io/projected/90536c4e-c461-4d2c-95dd-f08660cb2e69-kube-api-access-8q2gp\") pod \"auto-csr-approver-29567840-lm4bc\" (UID: \"90536c4e-c461-4d2c-95dd-f08660cb2e69\") " pod="openshift-infra/auto-csr-approver-29567840-lm4bc" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.485801 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52"] Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.487100 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.490456 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567840-lm4bc" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.490728 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.490805 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.494959 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.497058 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljtt5" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.525237 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52"] Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.615171 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8cf3d7cc-425b-4d40-a26c-a88d2b210c0a-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wtk52\" (UID: \"8cf3d7cc-425b-4d40-a26c-a88d2b210c0a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.615264 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxmhw\" (UniqueName: \"kubernetes.io/projected/8cf3d7cc-425b-4d40-a26c-a88d2b210c0a-kube-api-access-xxmhw\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wtk52\" (UID: \"8cf3d7cc-425b-4d40-a26c-a88d2b210c0a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.615491 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8cf3d7cc-425b-4d40-a26c-a88d2b210c0a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wtk52\" (UID: \"8cf3d7cc-425b-4d40-a26c-a88d2b210c0a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.720056 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8cf3d7cc-425b-4d40-a26c-a88d2b210c0a-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wtk52\" (UID: \"8cf3d7cc-425b-4d40-a26c-a88d2b210c0a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.720136 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxmhw\" (UniqueName: \"kubernetes.io/projected/8cf3d7cc-425b-4d40-a26c-a88d2b210c0a-kube-api-access-xxmhw\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wtk52\" (UID: \"8cf3d7cc-425b-4d40-a26c-a88d2b210c0a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.720551 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8cf3d7cc-425b-4d40-a26c-a88d2b210c0a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wtk52\" (UID: \"8cf3d7cc-425b-4d40-a26c-a88d2b210c0a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.726064 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8cf3d7cc-425b-4d40-a26c-a88d2b210c0a-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wtk52\" (UID: \"8cf3d7cc-425b-4d40-a26c-a88d2b210c0a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.727943 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8cf3d7cc-425b-4d40-a26c-a88d2b210c0a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wtk52\" (UID: \"8cf3d7cc-425b-4d40-a26c-a88d2b210c0a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.740853 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxmhw\" (UniqueName: \"kubernetes.io/projected/8cf3d7cc-425b-4d40-a26c-a88d2b210c0a-kube-api-access-xxmhw\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wtk52\" (UID: \"8cf3d7cc-425b-4d40-a26c-a88d2b210c0a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52" Mar 21 05:20:00 crc kubenswrapper[4775]: I0321 05:20:00.828020 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52" Mar 21 05:20:01 crc kubenswrapper[4775]: I0321 05:20:01.056637 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567840-lm4bc"] Mar 21 05:20:01 crc kubenswrapper[4775]: I0321 05:20:01.432441 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567840-lm4bc" event={"ID":"90536c4e-c461-4d2c-95dd-f08660cb2e69","Type":"ContainerStarted","Data":"d8db51af6d44c3f83810d412d859fe0105e96ac93de4bd29c22806706169bf8e"} Mar 21 05:20:01 crc kubenswrapper[4775]: I0321 05:20:01.484053 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52"] Mar 21 05:20:01 crc kubenswrapper[4775]: W0321 05:20:01.486271 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cf3d7cc_425b_4d40_a26c_a88d2b210c0a.slice/crio-6e010efa4efc7a62ef4d3be3e4b05434c3faa994dfeeba4e2ae97f91893d553e WatchSource:0}: Error finding container 6e010efa4efc7a62ef4d3be3e4b05434c3faa994dfeeba4e2ae97f91893d553e: Status 404 returned error can't find the container with id 6e010efa4efc7a62ef4d3be3e4b05434c3faa994dfeeba4e2ae97f91893d553e Mar 21 05:20:02 crc kubenswrapper[4775]: I0321 05:20:02.450278 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52" event={"ID":"8cf3d7cc-425b-4d40-a26c-a88d2b210c0a","Type":"ContainerStarted","Data":"7d776ce17e88d2b02f97757ee994a3ce5c88d09fa2bf2dab0d0c6611a107f314"} Mar 21 05:20:02 crc kubenswrapper[4775]: I0321 05:20:02.450811 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52" event={"ID":"8cf3d7cc-425b-4d40-a26c-a88d2b210c0a","Type":"ContainerStarted","Data":"6e010efa4efc7a62ef4d3be3e4b05434c3faa994dfeeba4e2ae97f91893d553e"} Mar 21 05:20:02 crc kubenswrapper[4775]: I0321 05:20:02.482363 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52" podStartSLOduration=1.931591224 podStartE2EDuration="2.482342183s" podCreationTimestamp="2026-03-21 05:20:00 +0000 UTC" firstStartedPulling="2026-03-21 05:20:01.48951521 +0000 UTC m=+1954.465978834" lastFinishedPulling="2026-03-21 05:20:02.040266169 +0000 UTC m=+1955.016729793" observedRunningTime="2026-03-21 05:20:02.47688896 +0000 UTC m=+1955.453352604" watchObservedRunningTime="2026-03-21 05:20:02.482342183 +0000 UTC m=+1955.458805807" Mar 21 05:20:03 crc kubenswrapper[4775]: I0321 05:20:03.470189 4775 generic.go:334] "Generic (PLEG): container finished" podID="90536c4e-c461-4d2c-95dd-f08660cb2e69" containerID="383196299bf8ab4c199e90d21197235aa0a2587594bef58700d15b41f3c4abec" exitCode=0 Mar 21 05:20:03 crc kubenswrapper[4775]: I0321 05:20:03.470313 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567840-lm4bc" event={"ID":"90536c4e-c461-4d2c-95dd-f08660cb2e69","Type":"ContainerDied","Data":"383196299bf8ab4c199e90d21197235aa0a2587594bef58700d15b41f3c4abec"} Mar 21 05:20:04 crc kubenswrapper[4775]: I0321 05:20:04.054457 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-s5ct6"] Mar 21 05:20:04 crc kubenswrapper[4775]: I0321 05:20:04.063242 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-s5ct6"] Mar 21 05:20:04 crc kubenswrapper[4775]: I0321 05:20:04.884251 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567840-lm4bc" Mar 21 05:20:05 crc kubenswrapper[4775]: I0321 05:20:05.106832 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q2gp\" (UniqueName: \"kubernetes.io/projected/90536c4e-c461-4d2c-95dd-f08660cb2e69-kube-api-access-8q2gp\") pod \"90536c4e-c461-4d2c-95dd-f08660cb2e69\" (UID: \"90536c4e-c461-4d2c-95dd-f08660cb2e69\") " Mar 21 05:20:05 crc kubenswrapper[4775]: I0321 05:20:05.115715 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90536c4e-c461-4d2c-95dd-f08660cb2e69-kube-api-access-8q2gp" (OuterVolumeSpecName: "kube-api-access-8q2gp") pod "90536c4e-c461-4d2c-95dd-f08660cb2e69" (UID: "90536c4e-c461-4d2c-95dd-f08660cb2e69"). InnerVolumeSpecName "kube-api-access-8q2gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:20:05 crc kubenswrapper[4775]: I0321 05:20:05.210010 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q2gp\" (UniqueName: \"kubernetes.io/projected/90536c4e-c461-4d2c-95dd-f08660cb2e69-kube-api-access-8q2gp\") on node \"crc\" DevicePath \"\"" Mar 21 05:20:05 crc kubenswrapper[4775]: I0321 05:20:05.499618 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567840-lm4bc" event={"ID":"90536c4e-c461-4d2c-95dd-f08660cb2e69","Type":"ContainerDied","Data":"d8db51af6d44c3f83810d412d859fe0105e96ac93de4bd29c22806706169bf8e"} Mar 21 05:20:05 crc kubenswrapper[4775]: I0321 05:20:05.499687 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8db51af6d44c3f83810d412d859fe0105e96ac93de4bd29c22806706169bf8e" Mar 21 05:20:05 crc kubenswrapper[4775]: I0321 05:20:05.499769 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567840-lm4bc" Mar 21 05:20:05 crc kubenswrapper[4775]: I0321 05:20:05.675589 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3172d76-da59-4dee-95df-6b74b7fb0033" path="/var/lib/kubelet/pods/f3172d76-da59-4dee-95df-6b74b7fb0033/volumes" Mar 21 05:20:05 crc kubenswrapper[4775]: I0321 05:20:05.961488 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567834-bqpk6"] Mar 21 05:20:05 crc kubenswrapper[4775]: I0321 05:20:05.971214 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567834-bqpk6"] Mar 21 05:20:07 crc kubenswrapper[4775]: I0321 05:20:07.678819 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dca04e19-ebb7-47c3-a588-e3dc932857bf" path="/var/lib/kubelet/pods/dca04e19-ebb7-47c3-a588-e3dc932857bf/volumes" Mar 21 05:20:11 crc kubenswrapper[4775]: I0321 05:20:11.615699 4775 generic.go:334] "Generic (PLEG): container finished" podID="8cf3d7cc-425b-4d40-a26c-a88d2b210c0a" containerID="7d776ce17e88d2b02f97757ee994a3ce5c88d09fa2bf2dab0d0c6611a107f314" exitCode=0 Mar 21 05:20:11 crc kubenswrapper[4775]: I0321 05:20:11.615803 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52" event={"ID":"8cf3d7cc-425b-4d40-a26c-a88d2b210c0a","Type":"ContainerDied","Data":"7d776ce17e88d2b02f97757ee994a3ce5c88d09fa2bf2dab0d0c6611a107f314"} Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.145504 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.261136 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8cf3d7cc-425b-4d40-a26c-a88d2b210c0a-inventory\") pod \"8cf3d7cc-425b-4d40-a26c-a88d2b210c0a\" (UID: \"8cf3d7cc-425b-4d40-a26c-a88d2b210c0a\") " Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.261254 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxmhw\" (UniqueName: \"kubernetes.io/projected/8cf3d7cc-425b-4d40-a26c-a88d2b210c0a-kube-api-access-xxmhw\") pod \"8cf3d7cc-425b-4d40-a26c-a88d2b210c0a\" (UID: \"8cf3d7cc-425b-4d40-a26c-a88d2b210c0a\") " Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.261694 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8cf3d7cc-425b-4d40-a26c-a88d2b210c0a-ssh-key-openstack-edpm-ipam\") pod \"8cf3d7cc-425b-4d40-a26c-a88d2b210c0a\" (UID: \"8cf3d7cc-425b-4d40-a26c-a88d2b210c0a\") " Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.269453 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cf3d7cc-425b-4d40-a26c-a88d2b210c0a-kube-api-access-xxmhw" (OuterVolumeSpecName: "kube-api-access-xxmhw") pod "8cf3d7cc-425b-4d40-a26c-a88d2b210c0a" (UID: "8cf3d7cc-425b-4d40-a26c-a88d2b210c0a"). InnerVolumeSpecName "kube-api-access-xxmhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.297676 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cf3d7cc-425b-4d40-a26c-a88d2b210c0a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8cf3d7cc-425b-4d40-a26c-a88d2b210c0a" (UID: "8cf3d7cc-425b-4d40-a26c-a88d2b210c0a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.298199 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cf3d7cc-425b-4d40-a26c-a88d2b210c0a-inventory" (OuterVolumeSpecName: "inventory") pod "8cf3d7cc-425b-4d40-a26c-a88d2b210c0a" (UID: "8cf3d7cc-425b-4d40-a26c-a88d2b210c0a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.364052 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8cf3d7cc-425b-4d40-a26c-a88d2b210c0a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.364109 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8cf3d7cc-425b-4d40-a26c-a88d2b210c0a-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.364146 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxmhw\" (UniqueName: \"kubernetes.io/projected/8cf3d7cc-425b-4d40-a26c-a88d2b210c0a-kube-api-access-xxmhw\") on node \"crc\" DevicePath \"\"" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.663567 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.733516 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wtk52" event={"ID":"8cf3d7cc-425b-4d40-a26c-a88d2b210c0a","Type":"ContainerDied","Data":"6e010efa4efc7a62ef4d3be3e4b05434c3faa994dfeeba4e2ae97f91893d553e"} Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.734260 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e010efa4efc7a62ef4d3be3e4b05434c3faa994dfeeba4e2ae97f91893d553e" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.749508 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6"] Mar 21 05:20:13 crc kubenswrapper[4775]: E0321 05:20:13.751185 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90536c4e-c461-4d2c-95dd-f08660cb2e69" containerName="oc" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.751205 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="90536c4e-c461-4d2c-95dd-f08660cb2e69" containerName="oc" Mar 21 05:20:13 crc kubenswrapper[4775]: E0321 05:20:13.751229 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cf3d7cc-425b-4d40-a26c-a88d2b210c0a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.751237 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cf3d7cc-425b-4d40-a26c-a88d2b210c0a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.751459 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="90536c4e-c461-4d2c-95dd-f08660cb2e69" containerName="oc" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.751470 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cf3d7cc-425b-4d40-a26c-a88d2b210c0a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.752177 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.762233 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.762530 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.762678 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljtt5" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.762780 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.779045 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6"] Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.876212 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/034f630d-d6d6-41f0-8df6-e5db37b778f3-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6\" (UID: \"034f630d-d6d6-41f0-8df6-e5db37b778f3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.876476 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm6hn\" (UniqueName: \"kubernetes.io/projected/034f630d-d6d6-41f0-8df6-e5db37b778f3-kube-api-access-fm6hn\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6\" (UID: \"034f630d-d6d6-41f0-8df6-e5db37b778f3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.876609 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/034f630d-d6d6-41f0-8df6-e5db37b778f3-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6\" (UID: \"034f630d-d6d6-41f0-8df6-e5db37b778f3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.978928 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm6hn\" (UniqueName: \"kubernetes.io/projected/034f630d-d6d6-41f0-8df6-e5db37b778f3-kube-api-access-fm6hn\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6\" (UID: \"034f630d-d6d6-41f0-8df6-e5db37b778f3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.979025 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/034f630d-d6d6-41f0-8df6-e5db37b778f3-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6\" (UID: \"034f630d-d6d6-41f0-8df6-e5db37b778f3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.979108 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/034f630d-d6d6-41f0-8df6-e5db37b778f3-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6\" (UID: \"034f630d-d6d6-41f0-8df6-e5db37b778f3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.983945 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/034f630d-d6d6-41f0-8df6-e5db37b778f3-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6\" (UID: \"034f630d-d6d6-41f0-8df6-e5db37b778f3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6" Mar 21 05:20:13 crc kubenswrapper[4775]: I0321 05:20:13.985293 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/034f630d-d6d6-41f0-8df6-e5db37b778f3-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6\" (UID: \"034f630d-d6d6-41f0-8df6-e5db37b778f3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6" Mar 21 05:20:14 crc kubenswrapper[4775]: I0321 05:20:14.005461 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm6hn\" (UniqueName: \"kubernetes.io/projected/034f630d-d6d6-41f0-8df6-e5db37b778f3-kube-api-access-fm6hn\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6\" (UID: \"034f630d-d6d6-41f0-8df6-e5db37b778f3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6" Mar 21 05:20:14 crc kubenswrapper[4775]: I0321 05:20:14.084990 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6" Mar 21 05:20:14 crc kubenswrapper[4775]: I0321 05:20:14.660292 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6"] Mar 21 05:20:14 crc kubenswrapper[4775]: I0321 05:20:14.699625 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6" event={"ID":"034f630d-d6d6-41f0-8df6-e5db37b778f3","Type":"ContainerStarted","Data":"5f2a16a34435a3935460d73dfabdb86ed489cb9cf19fa06ddc502da14005e057"} Mar 21 05:20:17 crc kubenswrapper[4775]: I0321 05:20:17.758972 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6" event={"ID":"034f630d-d6d6-41f0-8df6-e5db37b778f3","Type":"ContainerStarted","Data":"fe3a52e25338bc80c572fe45a80d5fa4e0c87be67de390c6641af69c0274c9cc"} Mar 21 05:20:17 crc kubenswrapper[4775]: I0321 05:20:17.782699 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6" podStartSLOduration=2.717090851 podStartE2EDuration="4.782679084s" podCreationTimestamp="2026-03-21 05:20:13 +0000 UTC" firstStartedPulling="2026-03-21 05:20:14.675640231 +0000 UTC m=+1967.652103855" lastFinishedPulling="2026-03-21 05:20:16.741228464 +0000 UTC m=+1969.717692088" observedRunningTime="2026-03-21 05:20:17.775319847 +0000 UTC m=+1970.751783471" watchObservedRunningTime="2026-03-21 05:20:17.782679084 +0000 UTC m=+1970.759142708" Mar 21 05:20:27 crc kubenswrapper[4775]: I0321 05:20:27.854636 4775 generic.go:334] "Generic (PLEG): container finished" podID="034f630d-d6d6-41f0-8df6-e5db37b778f3" containerID="fe3a52e25338bc80c572fe45a80d5fa4e0c87be67de390c6641af69c0274c9cc" exitCode=0 Mar 21 05:20:27 crc kubenswrapper[4775]: I0321 05:20:27.854712 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6" event={"ID":"034f630d-d6d6-41f0-8df6-e5db37b778f3","Type":"ContainerDied","Data":"fe3a52e25338bc80c572fe45a80d5fa4e0c87be67de390c6641af69c0274c9cc"} Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.343486 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6" Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.456696 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/034f630d-d6d6-41f0-8df6-e5db37b778f3-inventory\") pod \"034f630d-d6d6-41f0-8df6-e5db37b778f3\" (UID: \"034f630d-d6d6-41f0-8df6-e5db37b778f3\") " Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.456861 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/034f630d-d6d6-41f0-8df6-e5db37b778f3-ssh-key-openstack-edpm-ipam\") pod \"034f630d-d6d6-41f0-8df6-e5db37b778f3\" (UID: \"034f630d-d6d6-41f0-8df6-e5db37b778f3\") " Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.457062 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fm6hn\" (UniqueName: \"kubernetes.io/projected/034f630d-d6d6-41f0-8df6-e5db37b778f3-kube-api-access-fm6hn\") pod \"034f630d-d6d6-41f0-8df6-e5db37b778f3\" (UID: \"034f630d-d6d6-41f0-8df6-e5db37b778f3\") " Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.464405 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/034f630d-d6d6-41f0-8df6-e5db37b778f3-kube-api-access-fm6hn" (OuterVolumeSpecName: "kube-api-access-fm6hn") pod "034f630d-d6d6-41f0-8df6-e5db37b778f3" (UID: "034f630d-d6d6-41f0-8df6-e5db37b778f3"). InnerVolumeSpecName "kube-api-access-fm6hn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.491986 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/034f630d-d6d6-41f0-8df6-e5db37b778f3-inventory" (OuterVolumeSpecName: "inventory") pod "034f630d-d6d6-41f0-8df6-e5db37b778f3" (UID: "034f630d-d6d6-41f0-8df6-e5db37b778f3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.493965 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/034f630d-d6d6-41f0-8df6-e5db37b778f3-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "034f630d-d6d6-41f0-8df6-e5db37b778f3" (UID: "034f630d-d6d6-41f0-8df6-e5db37b778f3"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.560351 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fm6hn\" (UniqueName: \"kubernetes.io/projected/034f630d-d6d6-41f0-8df6-e5db37b778f3-kube-api-access-fm6hn\") on node \"crc\" DevicePath \"\"" Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.560413 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/034f630d-d6d6-41f0-8df6-e5db37b778f3-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.560425 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/034f630d-d6d6-41f0-8df6-e5db37b778f3-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.892568 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6" event={"ID":"034f630d-d6d6-41f0-8df6-e5db37b778f3","Type":"ContainerDied","Data":"5f2a16a34435a3935460d73dfabdb86ed489cb9cf19fa06ddc502da14005e057"} Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.892624 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f2a16a34435a3935460d73dfabdb86ed489cb9cf19fa06ddc502da14005e057" Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.892653 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6" Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.982839 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f"] Mar 21 05:20:29 crc kubenswrapper[4775]: E0321 05:20:29.983413 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="034f630d-d6d6-41f0-8df6-e5db37b778f3" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.983443 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="034f630d-d6d6-41f0-8df6-e5db37b778f3" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.983684 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="034f630d-d6d6-41f0-8df6-e5db37b778f3" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.984731 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.990516 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.990826 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.994166 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.994442 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.995666 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.995763 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljtt5" Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.995985 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 05:20:29 crc kubenswrapper[4775]: I0321 05:20:29.998269 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.002353 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f"] Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.078415 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.078500 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.078527 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.078557 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.078612 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.078831 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.079042 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.079112 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.079310 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.079537 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.079658 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z46m9\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-kube-api-access-z46m9\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.079714 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.079835 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.079875 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.182350 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.182454 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.182516 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.182557 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.182603 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.182694 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.182735 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z46m9\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-kube-api-access-z46m9\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.182774 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.182836 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.182874 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.182950 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.183012 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.183038 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.183078 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.187556 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.188586 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.188772 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.189605 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.189637 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.189675 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.191075 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.191943 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.192676 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.192774 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.193479 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.200112 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.200960 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.206036 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z46m9\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-kube-api-access-z46m9\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.303214 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.855901 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f"] Mar 21 05:20:30 crc kubenswrapper[4775]: W0321 05:20:30.870086 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8390751b_3911_4a24_a1a2_c3d1d10da875.slice/crio-210bdffc220ee9aa5fdb6ed382504df48708beccdf4e3e885144cf89e9272f6c WatchSource:0}: Error finding container 210bdffc220ee9aa5fdb6ed382504df48708beccdf4e3e885144cf89e9272f6c: Status 404 returned error can't find the container with id 210bdffc220ee9aa5fdb6ed382504df48708beccdf4e3e885144cf89e9272f6c Mar 21 05:20:30 crc kubenswrapper[4775]: I0321 05:20:30.903935 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" event={"ID":"8390751b-3911-4a24-a1a2-c3d1d10da875","Type":"ContainerStarted","Data":"210bdffc220ee9aa5fdb6ed382504df48708beccdf4e3e885144cf89e9272f6c"} Mar 21 05:20:32 crc kubenswrapper[4775]: I0321 05:20:32.923759 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" event={"ID":"8390751b-3911-4a24-a1a2-c3d1d10da875","Type":"ContainerStarted","Data":"fb4a533700f8e88d08d3b8894d316fb781c0daefbc69e4bdf3fb51c50ad22ac6"} Mar 21 05:20:32 crc kubenswrapper[4775]: I0321 05:20:32.953225 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" podStartSLOduration=3.148408717 podStartE2EDuration="3.953198276s" podCreationTimestamp="2026-03-21 05:20:29 +0000 UTC" firstStartedPulling="2026-03-21 05:20:30.873140877 +0000 UTC m=+1983.849604501" lastFinishedPulling="2026-03-21 05:20:31.677930436 +0000 UTC m=+1984.654394060" observedRunningTime="2026-03-21 05:20:32.941213949 +0000 UTC m=+1985.917677603" watchObservedRunningTime="2026-03-21 05:20:32.953198276 +0000 UTC m=+1985.929661900" Mar 21 05:20:46 crc kubenswrapper[4775]: I0321 05:20:46.721825 4775 scope.go:117] "RemoveContainer" containerID="a1c552cc0e8c345fb5a90c312288f579e557c98ae97907195961586729f792ec" Mar 21 05:20:46 crc kubenswrapper[4775]: I0321 05:20:46.841922 4775 scope.go:117] "RemoveContainer" containerID="244997576ccb04a6089eebd186997e7f409b2534428a82f578796006c7c674ed" Mar 21 05:21:07 crc kubenswrapper[4775]: I0321 05:21:07.256850 4775 generic.go:334] "Generic (PLEG): container finished" podID="8390751b-3911-4a24-a1a2-c3d1d10da875" containerID="fb4a533700f8e88d08d3b8894d316fb781c0daefbc69e4bdf3fb51c50ad22ac6" exitCode=0 Mar 21 05:21:07 crc kubenswrapper[4775]: I0321 05:21:07.256968 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" event={"ID":"8390751b-3911-4a24-a1a2-c3d1d10da875","Type":"ContainerDied","Data":"fb4a533700f8e88d08d3b8894d316fb781c0daefbc69e4bdf3fb51c50ad22ac6"} Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.778807 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.845868 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-ssh-key-openstack-edpm-ipam\") pod \"8390751b-3911-4a24-a1a2-c3d1d10da875\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.845922 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-repo-setup-combined-ca-bundle\") pod \"8390751b-3911-4a24-a1a2-c3d1d10da875\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.845980 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-telemetry-combined-ca-bundle\") pod \"8390751b-3911-4a24-a1a2-c3d1d10da875\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.846019 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-ovn-combined-ca-bundle\") pod \"8390751b-3911-4a24-a1a2-c3d1d10da875\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.846049 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-nova-combined-ca-bundle\") pod \"8390751b-3911-4a24-a1a2-c3d1d10da875\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.846088 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-libvirt-combined-ca-bundle\") pod \"8390751b-3911-4a24-a1a2-c3d1d10da875\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.846157 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"8390751b-3911-4a24-a1a2-c3d1d10da875\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.846208 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z46m9\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-kube-api-access-z46m9\") pod \"8390751b-3911-4a24-a1a2-c3d1d10da875\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.846267 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"8390751b-3911-4a24-a1a2-c3d1d10da875\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.846301 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-neutron-metadata-combined-ca-bundle\") pod \"8390751b-3911-4a24-a1a2-c3d1d10da875\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.846328 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-bootstrap-combined-ca-bundle\") pod \"8390751b-3911-4a24-a1a2-c3d1d10da875\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.846358 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-inventory\") pod \"8390751b-3911-4a24-a1a2-c3d1d10da875\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.846403 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-ovn-default-certs-0\") pod \"8390751b-3911-4a24-a1a2-c3d1d10da875\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.846431 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"8390751b-3911-4a24-a1a2-c3d1d10da875\" (UID: \"8390751b-3911-4a24-a1a2-c3d1d10da875\") " Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.857484 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-kube-api-access-z46m9" (OuterVolumeSpecName: "kube-api-access-z46m9") pod "8390751b-3911-4a24-a1a2-c3d1d10da875" (UID: "8390751b-3911-4a24-a1a2-c3d1d10da875"). InnerVolumeSpecName "kube-api-access-z46m9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.859822 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "8390751b-3911-4a24-a1a2-c3d1d10da875" (UID: "8390751b-3911-4a24-a1a2-c3d1d10da875"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.859909 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "8390751b-3911-4a24-a1a2-c3d1d10da875" (UID: "8390751b-3911-4a24-a1a2-c3d1d10da875"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.860695 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "8390751b-3911-4a24-a1a2-c3d1d10da875" (UID: "8390751b-3911-4a24-a1a2-c3d1d10da875"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.860741 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "8390751b-3911-4a24-a1a2-c3d1d10da875" (UID: "8390751b-3911-4a24-a1a2-c3d1d10da875"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.860863 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "8390751b-3911-4a24-a1a2-c3d1d10da875" (UID: "8390751b-3911-4a24-a1a2-c3d1d10da875"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.860949 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "8390751b-3911-4a24-a1a2-c3d1d10da875" (UID: "8390751b-3911-4a24-a1a2-c3d1d10da875"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.861004 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "8390751b-3911-4a24-a1a2-c3d1d10da875" (UID: "8390751b-3911-4a24-a1a2-c3d1d10da875"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.861022 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "8390751b-3911-4a24-a1a2-c3d1d10da875" (UID: "8390751b-3911-4a24-a1a2-c3d1d10da875"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.861568 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "8390751b-3911-4a24-a1a2-c3d1d10da875" (UID: "8390751b-3911-4a24-a1a2-c3d1d10da875"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.862370 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "8390751b-3911-4a24-a1a2-c3d1d10da875" (UID: "8390751b-3911-4a24-a1a2-c3d1d10da875"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.871476 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "8390751b-3911-4a24-a1a2-c3d1d10da875" (UID: "8390751b-3911-4a24-a1a2-c3d1d10da875"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.889297 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-inventory" (OuterVolumeSpecName: "inventory") pod "8390751b-3911-4a24-a1a2-c3d1d10da875" (UID: "8390751b-3911-4a24-a1a2-c3d1d10da875"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.889763 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8390751b-3911-4a24-a1a2-c3d1d10da875" (UID: "8390751b-3911-4a24-a1a2-c3d1d10da875"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.949191 4775 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.949244 4775 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.949258 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.949268 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.949280 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.949299 4775 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.949313 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.949324 4775 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.949336 4775 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.949345 4775 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.949355 4775 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8390751b-3911-4a24-a1a2-c3d1d10da875-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.949367 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.949378 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z46m9\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-kube-api-access-z46m9\") on node \"crc\" DevicePath \"\"" Mar 21 05:21:08 crc kubenswrapper[4775]: I0321 05:21:08.949388 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8390751b-3911-4a24-a1a2-c3d1d10da875-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.279325 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" event={"ID":"8390751b-3911-4a24-a1a2-c3d1d10da875","Type":"ContainerDied","Data":"210bdffc220ee9aa5fdb6ed382504df48708beccdf4e3e885144cf89e9272f6c"} Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.280033 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="210bdffc220ee9aa5fdb6ed382504df48708beccdf4e3e885144cf89e9272f6c" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.279434 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.536647 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj"] Mar 21 05:21:09 crc kubenswrapper[4775]: E0321 05:21:09.537219 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8390751b-3911-4a24-a1a2-c3d1d10da875" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.537239 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8390751b-3911-4a24-a1a2-c3d1d10da875" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.537469 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8390751b-3911-4a24-a1a2-c3d1d10da875" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.540238 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.544019 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.549305 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.552814 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljtt5" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.553065 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.553199 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.555166 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj"] Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.690910 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-45kkj\" (UID: \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.691015 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-45kkj\" (UID: \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.691065 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-45kkj\" (UID: \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.691352 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-45kkj\" (UID: \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.691488 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff8vp\" (UniqueName: \"kubernetes.io/projected/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-kube-api-access-ff8vp\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-45kkj\" (UID: \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.794263 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-45kkj\" (UID: \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.794359 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-45kkj\" (UID: \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.794386 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-45kkj\" (UID: \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.794560 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-45kkj\" (UID: \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.794647 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff8vp\" (UniqueName: \"kubernetes.io/projected/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-kube-api-access-ff8vp\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-45kkj\" (UID: \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.795299 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-45kkj\" (UID: \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.799857 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-45kkj\" (UID: \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.800911 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-45kkj\" (UID: \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.805787 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-45kkj\" (UID: \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.825252 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff8vp\" (UniqueName: \"kubernetes.io/projected/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-kube-api-access-ff8vp\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-45kkj\" (UID: \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" Mar 21 05:21:09 crc kubenswrapper[4775]: I0321 05:21:09.863296 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" Mar 21 05:21:10 crc kubenswrapper[4775]: I0321 05:21:10.442020 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj"] Mar 21 05:21:11 crc kubenswrapper[4775]: I0321 05:21:11.305369 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" event={"ID":"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d","Type":"ContainerStarted","Data":"91615bfc7851440e5b50d2e960a1c3c69c8e29ff7364666e05cee85e6021db4b"} Mar 21 05:21:13 crc kubenswrapper[4775]: I0321 05:21:13.330426 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" event={"ID":"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d","Type":"ContainerStarted","Data":"44f7e0cba4f9c13c945de5cfc1e82d38f03c6e9791b0edfa1200a4d76e980571"} Mar 21 05:21:13 crc kubenswrapper[4775]: I0321 05:21:13.357153 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" podStartSLOduration=2.5938170080000003 podStartE2EDuration="4.357093854s" podCreationTimestamp="2026-03-21 05:21:09 +0000 UTC" firstStartedPulling="2026-03-21 05:21:10.448626553 +0000 UTC m=+2023.425090177" lastFinishedPulling="2026-03-21 05:21:12.211903399 +0000 UTC m=+2025.188367023" observedRunningTime="2026-03-21 05:21:13.352169056 +0000 UTC m=+2026.328632690" watchObservedRunningTime="2026-03-21 05:21:13.357093854 +0000 UTC m=+2026.333557478" Mar 21 05:21:32 crc kubenswrapper[4775]: I0321 05:21:32.482266 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:21:32 crc kubenswrapper[4775]: I0321 05:21:32.482865 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:22:00 crc kubenswrapper[4775]: I0321 05:22:00.151497 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567842-zs7vw"] Mar 21 05:22:00 crc kubenswrapper[4775]: I0321 05:22:00.153889 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567842-zs7vw" Mar 21 05:22:00 crc kubenswrapper[4775]: I0321 05:22:00.158541 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:22:00 crc kubenswrapper[4775]: I0321 05:22:00.158789 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:22:00 crc kubenswrapper[4775]: I0321 05:22:00.161497 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:22:00 crc kubenswrapper[4775]: I0321 05:22:00.164682 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567842-zs7vw"] Mar 21 05:22:00 crc kubenswrapper[4775]: I0321 05:22:00.243941 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mxh4\" (UniqueName: \"kubernetes.io/projected/616ac3d7-a66e-45e3-b9ca-257dbd29e212-kube-api-access-7mxh4\") pod \"auto-csr-approver-29567842-zs7vw\" (UID: \"616ac3d7-a66e-45e3-b9ca-257dbd29e212\") " pod="openshift-infra/auto-csr-approver-29567842-zs7vw" Mar 21 05:22:00 crc kubenswrapper[4775]: I0321 05:22:00.346031 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mxh4\" (UniqueName: \"kubernetes.io/projected/616ac3d7-a66e-45e3-b9ca-257dbd29e212-kube-api-access-7mxh4\") pod \"auto-csr-approver-29567842-zs7vw\" (UID: \"616ac3d7-a66e-45e3-b9ca-257dbd29e212\") " pod="openshift-infra/auto-csr-approver-29567842-zs7vw" Mar 21 05:22:00 crc kubenswrapper[4775]: I0321 05:22:00.368397 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mxh4\" (UniqueName: \"kubernetes.io/projected/616ac3d7-a66e-45e3-b9ca-257dbd29e212-kube-api-access-7mxh4\") pod \"auto-csr-approver-29567842-zs7vw\" (UID: \"616ac3d7-a66e-45e3-b9ca-257dbd29e212\") " pod="openshift-infra/auto-csr-approver-29567842-zs7vw" Mar 21 05:22:00 crc kubenswrapper[4775]: I0321 05:22:00.474262 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567842-zs7vw" Mar 21 05:22:00 crc kubenswrapper[4775]: I0321 05:22:00.930590 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567842-zs7vw"] Mar 21 05:22:01 crc kubenswrapper[4775]: I0321 05:22:01.812578 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567842-zs7vw" event={"ID":"616ac3d7-a66e-45e3-b9ca-257dbd29e212","Type":"ContainerStarted","Data":"468537eb5adb84885c64524865c1e78dfcc317bb1fa504df59db45e3a6ec9df0"} Mar 21 05:22:02 crc kubenswrapper[4775]: I0321 05:22:02.482364 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:22:02 crc kubenswrapper[4775]: I0321 05:22:02.482766 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:22:02 crc kubenswrapper[4775]: I0321 05:22:02.824548 4775 generic.go:334] "Generic (PLEG): container finished" podID="616ac3d7-a66e-45e3-b9ca-257dbd29e212" containerID="a383e6d9b0bff1f3c71589e15f0f0c78443f3a1f70c9b149974bd8327eb1f252" exitCode=0 Mar 21 05:22:02 crc kubenswrapper[4775]: I0321 05:22:02.824595 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567842-zs7vw" event={"ID":"616ac3d7-a66e-45e3-b9ca-257dbd29e212","Type":"ContainerDied","Data":"a383e6d9b0bff1f3c71589e15f0f0c78443f3a1f70c9b149974bd8327eb1f252"} Mar 21 05:22:04 crc kubenswrapper[4775]: I0321 05:22:04.199894 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567842-zs7vw" Mar 21 05:22:04 crc kubenswrapper[4775]: I0321 05:22:04.348797 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mxh4\" (UniqueName: \"kubernetes.io/projected/616ac3d7-a66e-45e3-b9ca-257dbd29e212-kube-api-access-7mxh4\") pod \"616ac3d7-a66e-45e3-b9ca-257dbd29e212\" (UID: \"616ac3d7-a66e-45e3-b9ca-257dbd29e212\") " Mar 21 05:22:04 crc kubenswrapper[4775]: I0321 05:22:04.356083 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/616ac3d7-a66e-45e3-b9ca-257dbd29e212-kube-api-access-7mxh4" (OuterVolumeSpecName: "kube-api-access-7mxh4") pod "616ac3d7-a66e-45e3-b9ca-257dbd29e212" (UID: "616ac3d7-a66e-45e3-b9ca-257dbd29e212"). InnerVolumeSpecName "kube-api-access-7mxh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:22:04 crc kubenswrapper[4775]: I0321 05:22:04.452978 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mxh4\" (UniqueName: \"kubernetes.io/projected/616ac3d7-a66e-45e3-b9ca-257dbd29e212-kube-api-access-7mxh4\") on node \"crc\" DevicePath \"\"" Mar 21 05:22:04 crc kubenswrapper[4775]: I0321 05:22:04.843318 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567842-zs7vw" event={"ID":"616ac3d7-a66e-45e3-b9ca-257dbd29e212","Type":"ContainerDied","Data":"468537eb5adb84885c64524865c1e78dfcc317bb1fa504df59db45e3a6ec9df0"} Mar 21 05:22:04 crc kubenswrapper[4775]: I0321 05:22:04.843696 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="468537eb5adb84885c64524865c1e78dfcc317bb1fa504df59db45e3a6ec9df0" Mar 21 05:22:04 crc kubenswrapper[4775]: I0321 05:22:04.843371 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567842-zs7vw" Mar 21 05:22:05 crc kubenswrapper[4775]: I0321 05:22:05.288919 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567836-whhlh"] Mar 21 05:22:05 crc kubenswrapper[4775]: I0321 05:22:05.298088 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567836-whhlh"] Mar 21 05:22:05 crc kubenswrapper[4775]: I0321 05:22:05.673528 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc" path="/var/lib/kubelet/pods/4b08d38e-ec7f-4b0a-9f67-11a4b2b4d3fc/volumes" Mar 21 05:22:14 crc kubenswrapper[4775]: I0321 05:22:14.935168 4775 generic.go:334] "Generic (PLEG): container finished" podID="3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d" containerID="44f7e0cba4f9c13c945de5cfc1e82d38f03c6e9791b0edfa1200a4d76e980571" exitCode=0 Mar 21 05:22:14 crc kubenswrapper[4775]: I0321 05:22:14.935729 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" event={"ID":"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d","Type":"ContainerDied","Data":"44f7e0cba4f9c13c945de5cfc1e82d38f03c6e9791b0edfa1200a4d76e980571"} Mar 21 05:22:16 crc kubenswrapper[4775]: I0321 05:22:16.439294 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" Mar 21 05:22:16 crc kubenswrapper[4775]: I0321 05:22:16.502849 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ff8vp\" (UniqueName: \"kubernetes.io/projected/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-kube-api-access-ff8vp\") pod \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\" (UID: \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\") " Mar 21 05:22:16 crc kubenswrapper[4775]: I0321 05:22:16.502941 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-ssh-key-openstack-edpm-ipam\") pod \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\" (UID: \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\") " Mar 21 05:22:16 crc kubenswrapper[4775]: I0321 05:22:16.503288 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-ovn-combined-ca-bundle\") pod \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\" (UID: \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\") " Mar 21 05:22:16 crc kubenswrapper[4775]: I0321 05:22:16.503322 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-ovncontroller-config-0\") pod \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\" (UID: \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\") " Mar 21 05:22:16 crc kubenswrapper[4775]: I0321 05:22:16.503362 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-inventory\") pod \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\" (UID: \"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d\") " Mar 21 05:22:16 crc kubenswrapper[4775]: I0321 05:22:16.517396 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d" (UID: "3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:22:16 crc kubenswrapper[4775]: I0321 05:22:16.526395 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-kube-api-access-ff8vp" (OuterVolumeSpecName: "kube-api-access-ff8vp") pod "3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d" (UID: "3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d"). InnerVolumeSpecName "kube-api-access-ff8vp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:22:16 crc kubenswrapper[4775]: I0321 05:22:16.529707 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d" (UID: "3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:22:16 crc kubenswrapper[4775]: I0321 05:22:16.533876 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-inventory" (OuterVolumeSpecName: "inventory") pod "3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d" (UID: "3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:22:16 crc kubenswrapper[4775]: I0321 05:22:16.554222 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d" (UID: "3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:22:16 crc kubenswrapper[4775]: I0321 05:22:16.606300 4775 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:22:16 crc kubenswrapper[4775]: I0321 05:22:16.606354 4775 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:22:16 crc kubenswrapper[4775]: I0321 05:22:16.606370 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 05:22:16 crc kubenswrapper[4775]: I0321 05:22:16.606382 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ff8vp\" (UniqueName: \"kubernetes.io/projected/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-kube-api-access-ff8vp\") on node \"crc\" DevicePath \"\"" Mar 21 05:22:16 crc kubenswrapper[4775]: I0321 05:22:16.606393 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 05:22:16 crc kubenswrapper[4775]: I0321 05:22:16.959486 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" event={"ID":"3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d","Type":"ContainerDied","Data":"91615bfc7851440e5b50d2e960a1c3c69c8e29ff7364666e05cee85e6021db4b"} Mar 21 05:22:16 crc kubenswrapper[4775]: I0321 05:22:16.959540 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91615bfc7851440e5b50d2e960a1c3c69c8e29ff7364666e05cee85e6021db4b" Mar 21 05:22:16 crc kubenswrapper[4775]: I0321 05:22:16.959582 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-45kkj" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.140599 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9"] Mar 21 05:22:17 crc kubenswrapper[4775]: E0321 05:22:17.141181 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="616ac3d7-a66e-45e3-b9ca-257dbd29e212" containerName="oc" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.141207 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="616ac3d7-a66e-45e3-b9ca-257dbd29e212" containerName="oc" Mar 21 05:22:17 crc kubenswrapper[4775]: E0321 05:22:17.141238 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.141247 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.141465 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.141507 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="616ac3d7-a66e-45e3-b9ca-257dbd29e212" containerName="oc" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.142339 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.145256 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.145370 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.145399 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.145452 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.145565 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljtt5" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.145754 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.158106 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9"] Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.222352 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.222396 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6gcx\" (UniqueName: \"kubernetes.io/projected/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-kube-api-access-m6gcx\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.222508 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.222531 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.222567 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.222600 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.324915 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.325025 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.325798 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.325935 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.326178 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.326215 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6gcx\" (UniqueName: \"kubernetes.io/projected/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-kube-api-access-m6gcx\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.338382 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.338601 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.338637 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.338857 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.339510 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.351559 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6gcx\" (UniqueName: \"kubernetes.io/projected/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-kube-api-access-m6gcx\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.461654 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:22:17 crc kubenswrapper[4775]: I0321 05:22:17.986577 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9"] Mar 21 05:22:18 crc kubenswrapper[4775]: I0321 05:22:18.983464 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" event={"ID":"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70","Type":"ContainerStarted","Data":"a2c5ca368c5713780f99ace2b5deee2ccd41dd05d8fa3f6e617082e9fad7068e"} Mar 21 05:22:21 crc kubenswrapper[4775]: I0321 05:22:21.020187 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" event={"ID":"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70","Type":"ContainerStarted","Data":"b09ff33d04d4e01ab63f1e09911d9dd1b31d78a871761c9ff131b30c739eccb4"} Mar 21 05:22:21 crc kubenswrapper[4775]: I0321 05:22:21.052867 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" podStartSLOduration=2.212685062 podStartE2EDuration="4.052843338s" podCreationTimestamp="2026-03-21 05:22:17 +0000 UTC" firstStartedPulling="2026-03-21 05:22:17.99540236 +0000 UTC m=+2090.971865994" lastFinishedPulling="2026-03-21 05:22:19.835560636 +0000 UTC m=+2092.812024270" observedRunningTime="2026-03-21 05:22:21.037591209 +0000 UTC m=+2094.014054863" watchObservedRunningTime="2026-03-21 05:22:21.052843338 +0000 UTC m=+2094.029306962" Mar 21 05:22:32 crc kubenswrapper[4775]: I0321 05:22:32.482512 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:22:32 crc kubenswrapper[4775]: I0321 05:22:32.483081 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:22:32 crc kubenswrapper[4775]: I0321 05:22:32.483139 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 05:22:32 crc kubenswrapper[4775]: I0321 05:22:32.483636 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"820b5e7971de8f534813cc1fdb277aabd37bf9063cedc671ef2c92d0328150cb"} pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 05:22:32 crc kubenswrapper[4775]: I0321 05:22:32.483689 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" containerID="cri-o://820b5e7971de8f534813cc1fdb277aabd37bf9063cedc671ef2c92d0328150cb" gracePeriod=600 Mar 21 05:22:33 crc kubenswrapper[4775]: I0321 05:22:33.143154 4775 generic.go:334] "Generic (PLEG): container finished" podID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerID="820b5e7971de8f534813cc1fdb277aabd37bf9063cedc671ef2c92d0328150cb" exitCode=0 Mar 21 05:22:33 crc kubenswrapper[4775]: I0321 05:22:33.143215 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerDied","Data":"820b5e7971de8f534813cc1fdb277aabd37bf9063cedc671ef2c92d0328150cb"} Mar 21 05:22:33 crc kubenswrapper[4775]: I0321 05:22:33.143470 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerStarted","Data":"21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804"} Mar 21 05:22:33 crc kubenswrapper[4775]: I0321 05:22:33.143490 4775 scope.go:117] "RemoveContainer" containerID="5c991c4c4614039503bd13fbe36151a2388f655ed79ef2fa30df26614cdbcaa4" Mar 21 05:22:46 crc kubenswrapper[4775]: I0321 05:22:46.964568 4775 scope.go:117] "RemoveContainer" containerID="f116b6d526cd4aaa616f201bcf5084943e1e56754ce61bd36026bd8732dd3bd1" Mar 21 05:22:55 crc kubenswrapper[4775]: I0321 05:22:55.950390 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jtxhl"] Mar 21 05:22:55 crc kubenswrapper[4775]: I0321 05:22:55.954149 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jtxhl" Mar 21 05:22:55 crc kubenswrapper[4775]: I0321 05:22:55.972283 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jtxhl"] Mar 21 05:22:56 crc kubenswrapper[4775]: I0321 05:22:56.098047 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4871de6-5a5d-4c10-98d5-c73c21570947-utilities\") pod \"redhat-operators-jtxhl\" (UID: \"d4871de6-5a5d-4c10-98d5-c73c21570947\") " pod="openshift-marketplace/redhat-operators-jtxhl" Mar 21 05:22:56 crc kubenswrapper[4775]: I0321 05:22:56.098288 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2jvd\" (UniqueName: \"kubernetes.io/projected/d4871de6-5a5d-4c10-98d5-c73c21570947-kube-api-access-j2jvd\") pod \"redhat-operators-jtxhl\" (UID: \"d4871de6-5a5d-4c10-98d5-c73c21570947\") " pod="openshift-marketplace/redhat-operators-jtxhl" Mar 21 05:22:56 crc kubenswrapper[4775]: I0321 05:22:56.098335 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4871de6-5a5d-4c10-98d5-c73c21570947-catalog-content\") pod \"redhat-operators-jtxhl\" (UID: \"d4871de6-5a5d-4c10-98d5-c73c21570947\") " pod="openshift-marketplace/redhat-operators-jtxhl" Mar 21 05:22:56 crc kubenswrapper[4775]: I0321 05:22:56.200496 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2jvd\" (UniqueName: \"kubernetes.io/projected/d4871de6-5a5d-4c10-98d5-c73c21570947-kube-api-access-j2jvd\") pod \"redhat-operators-jtxhl\" (UID: \"d4871de6-5a5d-4c10-98d5-c73c21570947\") " pod="openshift-marketplace/redhat-operators-jtxhl" Mar 21 05:22:56 crc kubenswrapper[4775]: I0321 05:22:56.200986 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4871de6-5a5d-4c10-98d5-c73c21570947-catalog-content\") pod \"redhat-operators-jtxhl\" (UID: \"d4871de6-5a5d-4c10-98d5-c73c21570947\") " pod="openshift-marketplace/redhat-operators-jtxhl" Mar 21 05:22:56 crc kubenswrapper[4775]: I0321 05:22:56.201278 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4871de6-5a5d-4c10-98d5-c73c21570947-utilities\") pod \"redhat-operators-jtxhl\" (UID: \"d4871de6-5a5d-4c10-98d5-c73c21570947\") " pod="openshift-marketplace/redhat-operators-jtxhl" Mar 21 05:22:56 crc kubenswrapper[4775]: I0321 05:22:56.201627 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4871de6-5a5d-4c10-98d5-c73c21570947-catalog-content\") pod \"redhat-operators-jtxhl\" (UID: \"d4871de6-5a5d-4c10-98d5-c73c21570947\") " pod="openshift-marketplace/redhat-operators-jtxhl" Mar 21 05:22:56 crc kubenswrapper[4775]: I0321 05:22:56.202146 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4871de6-5a5d-4c10-98d5-c73c21570947-utilities\") pod \"redhat-operators-jtxhl\" (UID: \"d4871de6-5a5d-4c10-98d5-c73c21570947\") " pod="openshift-marketplace/redhat-operators-jtxhl" Mar 21 05:22:56 crc kubenswrapper[4775]: I0321 05:22:56.240687 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2jvd\" (UniqueName: \"kubernetes.io/projected/d4871de6-5a5d-4c10-98d5-c73c21570947-kube-api-access-j2jvd\") pod \"redhat-operators-jtxhl\" (UID: \"d4871de6-5a5d-4c10-98d5-c73c21570947\") " pod="openshift-marketplace/redhat-operators-jtxhl" Mar 21 05:22:56 crc kubenswrapper[4775]: I0321 05:22:56.317750 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jtxhl" Mar 21 05:22:56 crc kubenswrapper[4775]: I0321 05:22:56.882826 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jtxhl"] Mar 21 05:22:57 crc kubenswrapper[4775]: I0321 05:22:57.376286 4775 generic.go:334] "Generic (PLEG): container finished" podID="d4871de6-5a5d-4c10-98d5-c73c21570947" containerID="cd99f6278b5762bb26ce5bb130e6b5017494ab8ebb62ead08c30e6235d692bd9" exitCode=0 Mar 21 05:22:57 crc kubenswrapper[4775]: I0321 05:22:57.376364 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtxhl" event={"ID":"d4871de6-5a5d-4c10-98d5-c73c21570947","Type":"ContainerDied","Data":"cd99f6278b5762bb26ce5bb130e6b5017494ab8ebb62ead08c30e6235d692bd9"} Mar 21 05:22:57 crc kubenswrapper[4775]: I0321 05:22:57.376413 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtxhl" event={"ID":"d4871de6-5a5d-4c10-98d5-c73c21570947","Type":"ContainerStarted","Data":"133664e00450e69e13de7d7b6de59770930fe0b11d4ea3c3c34a4f85006bcffe"} Mar 21 05:22:59 crc kubenswrapper[4775]: I0321 05:22:59.398950 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtxhl" event={"ID":"d4871de6-5a5d-4c10-98d5-c73c21570947","Type":"ContainerStarted","Data":"441225cb122bf66cabf5c48f54264da9293c307b82c5b70551d4830941ce387b"} Mar 21 05:23:00 crc kubenswrapper[4775]: I0321 05:23:00.414423 4775 generic.go:334] "Generic (PLEG): container finished" podID="d4871de6-5a5d-4c10-98d5-c73c21570947" containerID="441225cb122bf66cabf5c48f54264da9293c307b82c5b70551d4830941ce387b" exitCode=0 Mar 21 05:23:00 crc kubenswrapper[4775]: I0321 05:23:00.414498 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtxhl" event={"ID":"d4871de6-5a5d-4c10-98d5-c73c21570947","Type":"ContainerDied","Data":"441225cb122bf66cabf5c48f54264da9293c307b82c5b70551d4830941ce387b"} Mar 21 05:23:01 crc kubenswrapper[4775]: I0321 05:23:01.428482 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtxhl" event={"ID":"d4871de6-5a5d-4c10-98d5-c73c21570947","Type":"ContainerStarted","Data":"a85357de462c6ddc32de73140ccff74411e8b1a6bb88a57b428ef54e3561339b"} Mar 21 05:23:01 crc kubenswrapper[4775]: I0321 05:23:01.458775 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jtxhl" podStartSLOduration=2.9898515679999997 podStartE2EDuration="6.458755721s" podCreationTimestamp="2026-03-21 05:22:55 +0000 UTC" firstStartedPulling="2026-03-21 05:22:57.379596397 +0000 UTC m=+2130.356060021" lastFinishedPulling="2026-03-21 05:23:00.84850056 +0000 UTC m=+2133.824964174" observedRunningTime="2026-03-21 05:23:01.450492009 +0000 UTC m=+2134.426955653" watchObservedRunningTime="2026-03-21 05:23:01.458755721 +0000 UTC m=+2134.435219345" Mar 21 05:23:06 crc kubenswrapper[4775]: I0321 05:23:06.318394 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jtxhl" Mar 21 05:23:06 crc kubenswrapper[4775]: I0321 05:23:06.319452 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jtxhl" Mar 21 05:23:07 crc kubenswrapper[4775]: I0321 05:23:07.376509 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jtxhl" podUID="d4871de6-5a5d-4c10-98d5-c73c21570947" containerName="registry-server" probeResult="failure" output=< Mar 21 05:23:07 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Mar 21 05:23:07 crc kubenswrapper[4775]: > Mar 21 05:23:09 crc kubenswrapper[4775]: I0321 05:23:09.508335 4775 generic.go:334] "Generic (PLEG): container finished" podID="29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70" containerID="b09ff33d04d4e01ab63f1e09911d9dd1b31d78a871761c9ff131b30c739eccb4" exitCode=0 Mar 21 05:23:09 crc kubenswrapper[4775]: I0321 05:23:09.508426 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" event={"ID":"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70","Type":"ContainerDied","Data":"b09ff33d04d4e01ab63f1e09911d9dd1b31d78a871761c9ff131b30c739eccb4"} Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.024276 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.131419 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-ssh-key-openstack-edpm-ipam\") pod \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.131488 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-nova-metadata-neutron-config-0\") pod \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.131572 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6gcx\" (UniqueName: \"kubernetes.io/projected/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-kube-api-access-m6gcx\") pod \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.131638 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-neutron-ovn-metadata-agent-neutron-config-0\") pod \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.131681 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-inventory\") pod \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.131770 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-neutron-metadata-combined-ca-bundle\") pod \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\" (UID: \"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70\") " Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.138577 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-kube-api-access-m6gcx" (OuterVolumeSpecName: "kube-api-access-m6gcx") pod "29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70" (UID: "29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70"). InnerVolumeSpecName "kube-api-access-m6gcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.149330 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70" (UID: "29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.166372 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70" (UID: "29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.168930 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70" (UID: "29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.172309 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70" (UID: "29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.173886 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-inventory" (OuterVolumeSpecName: "inventory") pod "29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70" (UID: "29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.234259 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.234302 4775 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.234313 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6gcx\" (UniqueName: \"kubernetes.io/projected/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-kube-api-access-m6gcx\") on node \"crc\" DevicePath \"\"" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.234325 4775 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.234337 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.234345 4775 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.530944 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" event={"ID":"29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70","Type":"ContainerDied","Data":"a2c5ca368c5713780f99ace2b5deee2ccd41dd05d8fa3f6e617082e9fad7068e"} Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.531364 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2c5ca368c5713780f99ace2b5deee2ccd41dd05d8fa3f6e617082e9fad7068e" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.531442 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.653267 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz"] Mar 21 05:23:11 crc kubenswrapper[4775]: E0321 05:23:11.653911 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.653944 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.654252 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.655226 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.658677 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.658735 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.658767 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.659487 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.659814 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljtt5" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.675017 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz"] Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.747454 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz\" (UID: \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.747598 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f844j\" (UniqueName: \"kubernetes.io/projected/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-kube-api-access-f844j\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz\" (UID: \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.747651 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz\" (UID: \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.747689 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz\" (UID: \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.747731 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz\" (UID: \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.849935 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz\" (UID: \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.850094 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f844j\" (UniqueName: \"kubernetes.io/projected/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-kube-api-access-f844j\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz\" (UID: \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.850165 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz\" (UID: \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.850204 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz\" (UID: \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.850244 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz\" (UID: \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.856228 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz\" (UID: \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.857343 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz\" (UID: \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.857609 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz\" (UID: \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.870870 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz\" (UID: \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.872374 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f844j\" (UniqueName: \"kubernetes.io/projected/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-kube-api-access-f844j\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz\" (UID: \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" Mar 21 05:23:11 crc kubenswrapper[4775]: I0321 05:23:11.978667 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" Mar 21 05:23:12 crc kubenswrapper[4775]: I0321 05:23:12.542303 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz"] Mar 21 05:23:13 crc kubenswrapper[4775]: I0321 05:23:13.552867 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" event={"ID":"a71dcc90-c70a-4ff8-bf4a-42f1a2415827","Type":"ContainerStarted","Data":"7a5a248239d97b47a9a8821636d4ceb0030545a6d3ce4d3885e064cebb074b6b"} Mar 21 05:23:14 crc kubenswrapper[4775]: I0321 05:23:14.566566 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" event={"ID":"a71dcc90-c70a-4ff8-bf4a-42f1a2415827","Type":"ContainerStarted","Data":"1e34cd56643c271492ee835c6208defe78742763b335409f4aab2ec44574dc78"} Mar 21 05:23:14 crc kubenswrapper[4775]: I0321 05:23:14.596877 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" podStartSLOduration=2.550810293 podStartE2EDuration="3.596841352s" podCreationTimestamp="2026-03-21 05:23:11 +0000 UTC" firstStartedPulling="2026-03-21 05:23:12.553615147 +0000 UTC m=+2145.530078771" lastFinishedPulling="2026-03-21 05:23:13.599646206 +0000 UTC m=+2146.576109830" observedRunningTime="2026-03-21 05:23:14.586056329 +0000 UTC m=+2147.562519993" watchObservedRunningTime="2026-03-21 05:23:14.596841352 +0000 UTC m=+2147.573305016" Mar 21 05:23:16 crc kubenswrapper[4775]: I0321 05:23:16.378277 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jtxhl" Mar 21 05:23:16 crc kubenswrapper[4775]: I0321 05:23:16.433307 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jtxhl" Mar 21 05:23:16 crc kubenswrapper[4775]: I0321 05:23:16.628055 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jtxhl"] Mar 21 05:23:17 crc kubenswrapper[4775]: I0321 05:23:17.600621 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jtxhl" podUID="d4871de6-5a5d-4c10-98d5-c73c21570947" containerName="registry-server" containerID="cri-o://a85357de462c6ddc32de73140ccff74411e8b1a6bb88a57b428ef54e3561339b" gracePeriod=2 Mar 21 05:23:18 crc kubenswrapper[4775]: I0321 05:23:18.619490 4775 generic.go:334] "Generic (PLEG): container finished" podID="d4871de6-5a5d-4c10-98d5-c73c21570947" containerID="a85357de462c6ddc32de73140ccff74411e8b1a6bb88a57b428ef54e3561339b" exitCode=0 Mar 21 05:23:18 crc kubenswrapper[4775]: I0321 05:23:18.619552 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtxhl" event={"ID":"d4871de6-5a5d-4c10-98d5-c73c21570947","Type":"ContainerDied","Data":"a85357de462c6ddc32de73140ccff74411e8b1a6bb88a57b428ef54e3561339b"} Mar 21 05:23:18 crc kubenswrapper[4775]: I0321 05:23:18.620342 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtxhl" event={"ID":"d4871de6-5a5d-4c10-98d5-c73c21570947","Type":"ContainerDied","Data":"133664e00450e69e13de7d7b6de59770930fe0b11d4ea3c3c34a4f85006bcffe"} Mar 21 05:23:18 crc kubenswrapper[4775]: I0321 05:23:18.620370 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="133664e00450e69e13de7d7b6de59770930fe0b11d4ea3c3c34a4f85006bcffe" Mar 21 05:23:18 crc kubenswrapper[4775]: I0321 05:23:18.626352 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jtxhl" Mar 21 05:23:18 crc kubenswrapper[4775]: I0321 05:23:18.719403 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4871de6-5a5d-4c10-98d5-c73c21570947-utilities\") pod \"d4871de6-5a5d-4c10-98d5-c73c21570947\" (UID: \"d4871de6-5a5d-4c10-98d5-c73c21570947\") " Mar 21 05:23:18 crc kubenswrapper[4775]: I0321 05:23:18.719686 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4871de6-5a5d-4c10-98d5-c73c21570947-catalog-content\") pod \"d4871de6-5a5d-4c10-98d5-c73c21570947\" (UID: \"d4871de6-5a5d-4c10-98d5-c73c21570947\") " Mar 21 05:23:18 crc kubenswrapper[4775]: I0321 05:23:18.719730 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2jvd\" (UniqueName: \"kubernetes.io/projected/d4871de6-5a5d-4c10-98d5-c73c21570947-kube-api-access-j2jvd\") pod \"d4871de6-5a5d-4c10-98d5-c73c21570947\" (UID: \"d4871de6-5a5d-4c10-98d5-c73c21570947\") " Mar 21 05:23:18 crc kubenswrapper[4775]: I0321 05:23:18.720968 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4871de6-5a5d-4c10-98d5-c73c21570947-utilities" (OuterVolumeSpecName: "utilities") pod "d4871de6-5a5d-4c10-98d5-c73c21570947" (UID: "d4871de6-5a5d-4c10-98d5-c73c21570947"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:23:18 crc kubenswrapper[4775]: I0321 05:23:18.727672 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4871de6-5a5d-4c10-98d5-c73c21570947-kube-api-access-j2jvd" (OuterVolumeSpecName: "kube-api-access-j2jvd") pod "d4871de6-5a5d-4c10-98d5-c73c21570947" (UID: "d4871de6-5a5d-4c10-98d5-c73c21570947"). InnerVolumeSpecName "kube-api-access-j2jvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:23:18 crc kubenswrapper[4775]: I0321 05:23:18.822566 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4871de6-5a5d-4c10-98d5-c73c21570947-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:23:18 crc kubenswrapper[4775]: I0321 05:23:18.822615 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2jvd\" (UniqueName: \"kubernetes.io/projected/d4871de6-5a5d-4c10-98d5-c73c21570947-kube-api-access-j2jvd\") on node \"crc\" DevicePath \"\"" Mar 21 05:23:18 crc kubenswrapper[4775]: I0321 05:23:18.853788 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4871de6-5a5d-4c10-98d5-c73c21570947-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4871de6-5a5d-4c10-98d5-c73c21570947" (UID: "d4871de6-5a5d-4c10-98d5-c73c21570947"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:23:18 crc kubenswrapper[4775]: I0321 05:23:18.924969 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4871de6-5a5d-4c10-98d5-c73c21570947-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:23:19 crc kubenswrapper[4775]: I0321 05:23:19.630785 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jtxhl" Mar 21 05:23:19 crc kubenswrapper[4775]: I0321 05:23:19.677675 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jtxhl"] Mar 21 05:23:19 crc kubenswrapper[4775]: I0321 05:23:19.684819 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jtxhl"] Mar 21 05:23:21 crc kubenswrapper[4775]: I0321 05:23:21.711779 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4871de6-5a5d-4c10-98d5-c73c21570947" path="/var/lib/kubelet/pods/d4871de6-5a5d-4c10-98d5-c73c21570947/volumes" Mar 21 05:23:22 crc kubenswrapper[4775]: I0321 05:23:22.037201 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8rzc5"] Mar 21 05:23:22 crc kubenswrapper[4775]: E0321 05:23:22.037836 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4871de6-5a5d-4c10-98d5-c73c21570947" containerName="extract-utilities" Mar 21 05:23:22 crc kubenswrapper[4775]: I0321 05:23:22.037858 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4871de6-5a5d-4c10-98d5-c73c21570947" containerName="extract-utilities" Mar 21 05:23:22 crc kubenswrapper[4775]: E0321 05:23:22.037915 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4871de6-5a5d-4c10-98d5-c73c21570947" containerName="extract-content" Mar 21 05:23:22 crc kubenswrapper[4775]: I0321 05:23:22.037924 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4871de6-5a5d-4c10-98d5-c73c21570947" containerName="extract-content" Mar 21 05:23:22 crc kubenswrapper[4775]: E0321 05:23:22.037939 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4871de6-5a5d-4c10-98d5-c73c21570947" containerName="registry-server" Mar 21 05:23:22 crc kubenswrapper[4775]: I0321 05:23:22.037947 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4871de6-5a5d-4c10-98d5-c73c21570947" containerName="registry-server" Mar 21 05:23:22 crc kubenswrapper[4775]: I0321 05:23:22.038186 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4871de6-5a5d-4c10-98d5-c73c21570947" containerName="registry-server" Mar 21 05:23:22 crc kubenswrapper[4775]: I0321 05:23:22.040063 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rzc5" Mar 21 05:23:22 crc kubenswrapper[4775]: I0321 05:23:22.053504 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8rzc5"] Mar 21 05:23:22 crc kubenswrapper[4775]: I0321 05:23:22.115617 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg7xp\" (UniqueName: \"kubernetes.io/projected/f864469c-129d-4283-97c3-a5bf7c335a55-kube-api-access-bg7xp\") pod \"certified-operators-8rzc5\" (UID: \"f864469c-129d-4283-97c3-a5bf7c335a55\") " pod="openshift-marketplace/certified-operators-8rzc5" Mar 21 05:23:22 crc kubenswrapper[4775]: I0321 05:23:22.115690 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f864469c-129d-4283-97c3-a5bf7c335a55-utilities\") pod \"certified-operators-8rzc5\" (UID: \"f864469c-129d-4283-97c3-a5bf7c335a55\") " pod="openshift-marketplace/certified-operators-8rzc5" Mar 21 05:23:22 crc kubenswrapper[4775]: I0321 05:23:22.115737 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f864469c-129d-4283-97c3-a5bf7c335a55-catalog-content\") pod \"certified-operators-8rzc5\" (UID: \"f864469c-129d-4283-97c3-a5bf7c335a55\") " pod="openshift-marketplace/certified-operators-8rzc5" Mar 21 05:23:22 crc kubenswrapper[4775]: I0321 05:23:22.217814 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg7xp\" (UniqueName: \"kubernetes.io/projected/f864469c-129d-4283-97c3-a5bf7c335a55-kube-api-access-bg7xp\") pod \"certified-operators-8rzc5\" (UID: \"f864469c-129d-4283-97c3-a5bf7c335a55\") " pod="openshift-marketplace/certified-operators-8rzc5" Mar 21 05:23:22 crc kubenswrapper[4775]: I0321 05:23:22.218451 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f864469c-129d-4283-97c3-a5bf7c335a55-utilities\") pod \"certified-operators-8rzc5\" (UID: \"f864469c-129d-4283-97c3-a5bf7c335a55\") " pod="openshift-marketplace/certified-operators-8rzc5" Mar 21 05:23:22 crc kubenswrapper[4775]: I0321 05:23:22.218526 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f864469c-129d-4283-97c3-a5bf7c335a55-catalog-content\") pod \"certified-operators-8rzc5\" (UID: \"f864469c-129d-4283-97c3-a5bf7c335a55\") " pod="openshift-marketplace/certified-operators-8rzc5" Mar 21 05:23:22 crc kubenswrapper[4775]: I0321 05:23:22.219043 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f864469c-129d-4283-97c3-a5bf7c335a55-utilities\") pod \"certified-operators-8rzc5\" (UID: \"f864469c-129d-4283-97c3-a5bf7c335a55\") " pod="openshift-marketplace/certified-operators-8rzc5" Mar 21 05:23:22 crc kubenswrapper[4775]: I0321 05:23:22.219315 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f864469c-129d-4283-97c3-a5bf7c335a55-catalog-content\") pod \"certified-operators-8rzc5\" (UID: \"f864469c-129d-4283-97c3-a5bf7c335a55\") " pod="openshift-marketplace/certified-operators-8rzc5" Mar 21 05:23:22 crc kubenswrapper[4775]: I0321 05:23:22.240043 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg7xp\" (UniqueName: \"kubernetes.io/projected/f864469c-129d-4283-97c3-a5bf7c335a55-kube-api-access-bg7xp\") pod \"certified-operators-8rzc5\" (UID: \"f864469c-129d-4283-97c3-a5bf7c335a55\") " pod="openshift-marketplace/certified-operators-8rzc5" Mar 21 05:23:22 crc kubenswrapper[4775]: I0321 05:23:22.488810 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rzc5" Mar 21 05:23:23 crc kubenswrapper[4775]: I0321 05:23:23.041106 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8rzc5"] Mar 21 05:23:23 crc kubenswrapper[4775]: I0321 05:23:23.709603 4775 generic.go:334] "Generic (PLEG): container finished" podID="f864469c-129d-4283-97c3-a5bf7c335a55" containerID="3ddd3f273e635e03c0656239a822df1967d7a5fa543be70834322e9a3f10b304" exitCode=0 Mar 21 05:23:23 crc kubenswrapper[4775]: I0321 05:23:23.710219 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rzc5" event={"ID":"f864469c-129d-4283-97c3-a5bf7c335a55","Type":"ContainerDied","Data":"3ddd3f273e635e03c0656239a822df1967d7a5fa543be70834322e9a3f10b304"} Mar 21 05:23:23 crc kubenswrapper[4775]: I0321 05:23:23.710253 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rzc5" event={"ID":"f864469c-129d-4283-97c3-a5bf7c335a55","Type":"ContainerStarted","Data":"eb54b40877e070b8198b71a3e3ca8585434f5e074af09bb5c6291db5937e329b"} Mar 21 05:23:25 crc kubenswrapper[4775]: I0321 05:23:25.731063 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rzc5" event={"ID":"f864469c-129d-4283-97c3-a5bf7c335a55","Type":"ContainerStarted","Data":"81aedb7b0f6632b1c0988935f00df406ff8ec25de082d813684693afdec87a41"} Mar 21 05:23:26 crc kubenswrapper[4775]: I0321 05:23:26.742996 4775 generic.go:334] "Generic (PLEG): container finished" podID="f864469c-129d-4283-97c3-a5bf7c335a55" containerID="81aedb7b0f6632b1c0988935f00df406ff8ec25de082d813684693afdec87a41" exitCode=0 Mar 21 05:23:26 crc kubenswrapper[4775]: I0321 05:23:26.743099 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rzc5" event={"ID":"f864469c-129d-4283-97c3-a5bf7c335a55","Type":"ContainerDied","Data":"81aedb7b0f6632b1c0988935f00df406ff8ec25de082d813684693afdec87a41"} Mar 21 05:23:28 crc kubenswrapper[4775]: I0321 05:23:28.770812 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rzc5" event={"ID":"f864469c-129d-4283-97c3-a5bf7c335a55","Type":"ContainerStarted","Data":"7e9592852ea8dface72034178e5cbac515817e1921dd9f93e7ce91f809b38e5a"} Mar 21 05:23:28 crc kubenswrapper[4775]: I0321 05:23:28.799953 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8rzc5" podStartSLOduration=2.347730028 podStartE2EDuration="6.799924275s" podCreationTimestamp="2026-03-21 05:23:22 +0000 UTC" firstStartedPulling="2026-03-21 05:23:23.712998889 +0000 UTC m=+2156.689462513" lastFinishedPulling="2026-03-21 05:23:28.165193146 +0000 UTC m=+2161.141656760" observedRunningTime="2026-03-21 05:23:28.793550756 +0000 UTC m=+2161.770014430" watchObservedRunningTime="2026-03-21 05:23:28.799924275 +0000 UTC m=+2161.776387899" Mar 21 05:23:32 crc kubenswrapper[4775]: I0321 05:23:32.489167 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8rzc5" Mar 21 05:23:32 crc kubenswrapper[4775]: I0321 05:23:32.489832 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8rzc5" Mar 21 05:23:32 crc kubenswrapper[4775]: I0321 05:23:32.545414 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8rzc5" Mar 21 05:23:42 crc kubenswrapper[4775]: I0321 05:23:42.543444 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8rzc5" Mar 21 05:23:42 crc kubenswrapper[4775]: I0321 05:23:42.598980 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8rzc5"] Mar 21 05:23:42 crc kubenswrapper[4775]: I0321 05:23:42.927475 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8rzc5" podUID="f864469c-129d-4283-97c3-a5bf7c335a55" containerName="registry-server" containerID="cri-o://7e9592852ea8dface72034178e5cbac515817e1921dd9f93e7ce91f809b38e5a" gracePeriod=2 Mar 21 05:23:43 crc kubenswrapper[4775]: I0321 05:23:43.398922 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rzc5" Mar 21 05:23:43 crc kubenswrapper[4775]: I0321 05:23:43.452611 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bg7xp\" (UniqueName: \"kubernetes.io/projected/f864469c-129d-4283-97c3-a5bf7c335a55-kube-api-access-bg7xp\") pod \"f864469c-129d-4283-97c3-a5bf7c335a55\" (UID: \"f864469c-129d-4283-97c3-a5bf7c335a55\") " Mar 21 05:23:43 crc kubenswrapper[4775]: I0321 05:23:43.452787 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f864469c-129d-4283-97c3-a5bf7c335a55-catalog-content\") pod \"f864469c-129d-4283-97c3-a5bf7c335a55\" (UID: \"f864469c-129d-4283-97c3-a5bf7c335a55\") " Mar 21 05:23:43 crc kubenswrapper[4775]: I0321 05:23:43.452836 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f864469c-129d-4283-97c3-a5bf7c335a55-utilities\") pod \"f864469c-129d-4283-97c3-a5bf7c335a55\" (UID: \"f864469c-129d-4283-97c3-a5bf7c335a55\") " Mar 21 05:23:43 crc kubenswrapper[4775]: I0321 05:23:43.454028 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f864469c-129d-4283-97c3-a5bf7c335a55-utilities" (OuterVolumeSpecName: "utilities") pod "f864469c-129d-4283-97c3-a5bf7c335a55" (UID: "f864469c-129d-4283-97c3-a5bf7c335a55"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:23:43 crc kubenswrapper[4775]: I0321 05:23:43.464321 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f864469c-129d-4283-97c3-a5bf7c335a55-kube-api-access-bg7xp" (OuterVolumeSpecName: "kube-api-access-bg7xp") pod "f864469c-129d-4283-97c3-a5bf7c335a55" (UID: "f864469c-129d-4283-97c3-a5bf7c335a55"). InnerVolumeSpecName "kube-api-access-bg7xp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:23:43 crc kubenswrapper[4775]: I0321 05:23:43.513854 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f864469c-129d-4283-97c3-a5bf7c335a55-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f864469c-129d-4283-97c3-a5bf7c335a55" (UID: "f864469c-129d-4283-97c3-a5bf7c335a55"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:23:43 crc kubenswrapper[4775]: I0321 05:23:43.555701 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f864469c-129d-4283-97c3-a5bf7c335a55-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:23:43 crc kubenswrapper[4775]: I0321 05:23:43.555744 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bg7xp\" (UniqueName: \"kubernetes.io/projected/f864469c-129d-4283-97c3-a5bf7c335a55-kube-api-access-bg7xp\") on node \"crc\" DevicePath \"\"" Mar 21 05:23:43 crc kubenswrapper[4775]: I0321 05:23:43.555762 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f864469c-129d-4283-97c3-a5bf7c335a55-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:23:43 crc kubenswrapper[4775]: I0321 05:23:43.952550 4775 generic.go:334] "Generic (PLEG): container finished" podID="f864469c-129d-4283-97c3-a5bf7c335a55" containerID="7e9592852ea8dface72034178e5cbac515817e1921dd9f93e7ce91f809b38e5a" exitCode=0 Mar 21 05:23:43 crc kubenswrapper[4775]: I0321 05:23:43.952658 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rzc5" event={"ID":"f864469c-129d-4283-97c3-a5bf7c335a55","Type":"ContainerDied","Data":"7e9592852ea8dface72034178e5cbac515817e1921dd9f93e7ce91f809b38e5a"} Mar 21 05:23:43 crc kubenswrapper[4775]: I0321 05:23:43.952701 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rzc5" Mar 21 05:23:43 crc kubenswrapper[4775]: I0321 05:23:43.953203 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rzc5" event={"ID":"f864469c-129d-4283-97c3-a5bf7c335a55","Type":"ContainerDied","Data":"eb54b40877e070b8198b71a3e3ca8585434f5e074af09bb5c6291db5937e329b"} Mar 21 05:23:43 crc kubenswrapper[4775]: I0321 05:23:43.953245 4775 scope.go:117] "RemoveContainer" containerID="7e9592852ea8dface72034178e5cbac515817e1921dd9f93e7ce91f809b38e5a" Mar 21 05:23:43 crc kubenswrapper[4775]: I0321 05:23:43.987294 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8rzc5"] Mar 21 05:23:43 crc kubenswrapper[4775]: I0321 05:23:43.987504 4775 scope.go:117] "RemoveContainer" containerID="81aedb7b0f6632b1c0988935f00df406ff8ec25de082d813684693afdec87a41" Mar 21 05:23:43 crc kubenswrapper[4775]: I0321 05:23:43.997623 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8rzc5"] Mar 21 05:23:44 crc kubenswrapper[4775]: I0321 05:23:44.019997 4775 scope.go:117] "RemoveContainer" containerID="3ddd3f273e635e03c0656239a822df1967d7a5fa543be70834322e9a3f10b304" Mar 21 05:23:44 crc kubenswrapper[4775]: I0321 05:23:44.074099 4775 scope.go:117] "RemoveContainer" containerID="7e9592852ea8dface72034178e5cbac515817e1921dd9f93e7ce91f809b38e5a" Mar 21 05:23:44 crc kubenswrapper[4775]: E0321 05:23:44.074931 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e9592852ea8dface72034178e5cbac515817e1921dd9f93e7ce91f809b38e5a\": container with ID starting with 7e9592852ea8dface72034178e5cbac515817e1921dd9f93e7ce91f809b38e5a not found: ID does not exist" containerID="7e9592852ea8dface72034178e5cbac515817e1921dd9f93e7ce91f809b38e5a" Mar 21 05:23:44 crc kubenswrapper[4775]: I0321 05:23:44.074973 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e9592852ea8dface72034178e5cbac515817e1921dd9f93e7ce91f809b38e5a"} err="failed to get container status \"7e9592852ea8dface72034178e5cbac515817e1921dd9f93e7ce91f809b38e5a\": rpc error: code = NotFound desc = could not find container \"7e9592852ea8dface72034178e5cbac515817e1921dd9f93e7ce91f809b38e5a\": container with ID starting with 7e9592852ea8dface72034178e5cbac515817e1921dd9f93e7ce91f809b38e5a not found: ID does not exist" Mar 21 05:23:44 crc kubenswrapper[4775]: I0321 05:23:44.075002 4775 scope.go:117] "RemoveContainer" containerID="81aedb7b0f6632b1c0988935f00df406ff8ec25de082d813684693afdec87a41" Mar 21 05:23:44 crc kubenswrapper[4775]: E0321 05:23:44.075426 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81aedb7b0f6632b1c0988935f00df406ff8ec25de082d813684693afdec87a41\": container with ID starting with 81aedb7b0f6632b1c0988935f00df406ff8ec25de082d813684693afdec87a41 not found: ID does not exist" containerID="81aedb7b0f6632b1c0988935f00df406ff8ec25de082d813684693afdec87a41" Mar 21 05:23:44 crc kubenswrapper[4775]: I0321 05:23:44.075467 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81aedb7b0f6632b1c0988935f00df406ff8ec25de082d813684693afdec87a41"} err="failed to get container status \"81aedb7b0f6632b1c0988935f00df406ff8ec25de082d813684693afdec87a41\": rpc error: code = NotFound desc = could not find container \"81aedb7b0f6632b1c0988935f00df406ff8ec25de082d813684693afdec87a41\": container with ID starting with 81aedb7b0f6632b1c0988935f00df406ff8ec25de082d813684693afdec87a41 not found: ID does not exist" Mar 21 05:23:44 crc kubenswrapper[4775]: I0321 05:23:44.075494 4775 scope.go:117] "RemoveContainer" containerID="3ddd3f273e635e03c0656239a822df1967d7a5fa543be70834322e9a3f10b304" Mar 21 05:23:44 crc kubenswrapper[4775]: E0321 05:23:44.076271 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ddd3f273e635e03c0656239a822df1967d7a5fa543be70834322e9a3f10b304\": container with ID starting with 3ddd3f273e635e03c0656239a822df1967d7a5fa543be70834322e9a3f10b304 not found: ID does not exist" containerID="3ddd3f273e635e03c0656239a822df1967d7a5fa543be70834322e9a3f10b304" Mar 21 05:23:44 crc kubenswrapper[4775]: I0321 05:23:44.076309 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ddd3f273e635e03c0656239a822df1967d7a5fa543be70834322e9a3f10b304"} err="failed to get container status \"3ddd3f273e635e03c0656239a822df1967d7a5fa543be70834322e9a3f10b304\": rpc error: code = NotFound desc = could not find container \"3ddd3f273e635e03c0656239a822df1967d7a5fa543be70834322e9a3f10b304\": container with ID starting with 3ddd3f273e635e03c0656239a822df1967d7a5fa543be70834322e9a3f10b304 not found: ID does not exist" Mar 21 05:23:45 crc kubenswrapper[4775]: I0321 05:23:45.674610 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f864469c-129d-4283-97c3-a5bf7c335a55" path="/var/lib/kubelet/pods/f864469c-129d-4283-97c3-a5bf7c335a55/volumes" Mar 21 05:24:00 crc kubenswrapper[4775]: I0321 05:24:00.151069 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567844-25mvs"] Mar 21 05:24:00 crc kubenswrapper[4775]: E0321 05:24:00.152380 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f864469c-129d-4283-97c3-a5bf7c335a55" containerName="registry-server" Mar 21 05:24:00 crc kubenswrapper[4775]: I0321 05:24:00.152402 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f864469c-129d-4283-97c3-a5bf7c335a55" containerName="registry-server" Mar 21 05:24:00 crc kubenswrapper[4775]: E0321 05:24:00.152424 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f864469c-129d-4283-97c3-a5bf7c335a55" containerName="extract-content" Mar 21 05:24:00 crc kubenswrapper[4775]: I0321 05:24:00.152432 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f864469c-129d-4283-97c3-a5bf7c335a55" containerName="extract-content" Mar 21 05:24:00 crc kubenswrapper[4775]: E0321 05:24:00.152451 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f864469c-129d-4283-97c3-a5bf7c335a55" containerName="extract-utilities" Mar 21 05:24:00 crc kubenswrapper[4775]: I0321 05:24:00.152459 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f864469c-129d-4283-97c3-a5bf7c335a55" containerName="extract-utilities" Mar 21 05:24:00 crc kubenswrapper[4775]: I0321 05:24:00.152761 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f864469c-129d-4283-97c3-a5bf7c335a55" containerName="registry-server" Mar 21 05:24:00 crc kubenswrapper[4775]: I0321 05:24:00.153862 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567844-25mvs" Mar 21 05:24:00 crc kubenswrapper[4775]: I0321 05:24:00.156589 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:24:00 crc kubenswrapper[4775]: I0321 05:24:00.156999 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:24:00 crc kubenswrapper[4775]: I0321 05:24:00.157852 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:24:00 crc kubenswrapper[4775]: I0321 05:24:00.164040 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567844-25mvs"] Mar 21 05:24:00 crc kubenswrapper[4775]: I0321 05:24:00.255722 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xrmw\" (UniqueName: \"kubernetes.io/projected/c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6-kube-api-access-2xrmw\") pod \"auto-csr-approver-29567844-25mvs\" (UID: \"c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6\") " pod="openshift-infra/auto-csr-approver-29567844-25mvs" Mar 21 05:24:00 crc kubenswrapper[4775]: I0321 05:24:00.357661 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xrmw\" (UniqueName: \"kubernetes.io/projected/c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6-kube-api-access-2xrmw\") pod \"auto-csr-approver-29567844-25mvs\" (UID: \"c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6\") " pod="openshift-infra/auto-csr-approver-29567844-25mvs" Mar 21 05:24:00 crc kubenswrapper[4775]: I0321 05:24:00.387848 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xrmw\" (UniqueName: \"kubernetes.io/projected/c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6-kube-api-access-2xrmw\") pod \"auto-csr-approver-29567844-25mvs\" (UID: \"c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6\") " pod="openshift-infra/auto-csr-approver-29567844-25mvs" Mar 21 05:24:00 crc kubenswrapper[4775]: I0321 05:24:00.478612 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567844-25mvs" Mar 21 05:24:00 crc kubenswrapper[4775]: I0321 05:24:00.987991 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567844-25mvs"] Mar 21 05:24:01 crc kubenswrapper[4775]: I0321 05:24:01.123636 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567844-25mvs" event={"ID":"c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6","Type":"ContainerStarted","Data":"3112a924be6ecbb9a71c82c7867b14424d440c7cfce364b4203e2701aca683fb"} Mar 21 05:24:03 crc kubenswrapper[4775]: I0321 05:24:03.142419 4775 generic.go:334] "Generic (PLEG): container finished" podID="c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6" containerID="5936961592e0ba9da2bfbec1bc18d0df5f38a71ab49e8c90edaa7dea4ccb02ee" exitCode=0 Mar 21 05:24:03 crc kubenswrapper[4775]: I0321 05:24:03.142523 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567844-25mvs" event={"ID":"c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6","Type":"ContainerDied","Data":"5936961592e0ba9da2bfbec1bc18d0df5f38a71ab49e8c90edaa7dea4ccb02ee"} Mar 21 05:24:04 crc kubenswrapper[4775]: I0321 05:24:04.553469 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567844-25mvs" Mar 21 05:24:04 crc kubenswrapper[4775]: I0321 05:24:04.661579 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xrmw\" (UniqueName: \"kubernetes.io/projected/c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6-kube-api-access-2xrmw\") pod \"c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6\" (UID: \"c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6\") " Mar 21 05:24:04 crc kubenswrapper[4775]: I0321 05:24:04.668850 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6-kube-api-access-2xrmw" (OuterVolumeSpecName: "kube-api-access-2xrmw") pod "c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6" (UID: "c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6"). InnerVolumeSpecName "kube-api-access-2xrmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:24:04 crc kubenswrapper[4775]: I0321 05:24:04.764855 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xrmw\" (UniqueName: \"kubernetes.io/projected/c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6-kube-api-access-2xrmw\") on node \"crc\" DevicePath \"\"" Mar 21 05:24:05 crc kubenswrapper[4775]: I0321 05:24:05.163487 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567844-25mvs" event={"ID":"c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6","Type":"ContainerDied","Data":"3112a924be6ecbb9a71c82c7867b14424d440c7cfce364b4203e2701aca683fb"} Mar 21 05:24:05 crc kubenswrapper[4775]: I0321 05:24:05.163850 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3112a924be6ecbb9a71c82c7867b14424d440c7cfce364b4203e2701aca683fb" Mar 21 05:24:05 crc kubenswrapper[4775]: I0321 05:24:05.163573 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567844-25mvs" Mar 21 05:24:05 crc kubenswrapper[4775]: I0321 05:24:05.626623 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567838-kc6hr"] Mar 21 05:24:05 crc kubenswrapper[4775]: I0321 05:24:05.635465 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567838-kc6hr"] Mar 21 05:24:05 crc kubenswrapper[4775]: I0321 05:24:05.675881 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ed18838-ce8f-499e-a043-886b7e878eb3" path="/var/lib/kubelet/pods/8ed18838-ce8f-499e-a043-886b7e878eb3/volumes" Mar 21 05:24:32 crc kubenswrapper[4775]: I0321 05:24:32.482059 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:24:32 crc kubenswrapper[4775]: I0321 05:24:32.482720 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:24:46 crc kubenswrapper[4775]: I0321 05:24:46.492665 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-px5pg"] Mar 21 05:24:46 crc kubenswrapper[4775]: E0321 05:24:46.493620 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6" containerName="oc" Mar 21 05:24:46 crc kubenswrapper[4775]: I0321 05:24:46.493634 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6" containerName="oc" Mar 21 05:24:46 crc kubenswrapper[4775]: I0321 05:24:46.493809 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6" containerName="oc" Mar 21 05:24:46 crc kubenswrapper[4775]: I0321 05:24:46.495201 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-px5pg" Mar 21 05:24:46 crc kubenswrapper[4775]: I0321 05:24:46.510474 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-px5pg"] Mar 21 05:24:46 crc kubenswrapper[4775]: I0321 05:24:46.555708 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aad02c72-91c8-4ce2-b69d-915d9b0af7ec-utilities\") pod \"redhat-marketplace-px5pg\" (UID: \"aad02c72-91c8-4ce2-b69d-915d9b0af7ec\") " pod="openshift-marketplace/redhat-marketplace-px5pg" Mar 21 05:24:46 crc kubenswrapper[4775]: I0321 05:24:46.555852 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzk6x\" (UniqueName: \"kubernetes.io/projected/aad02c72-91c8-4ce2-b69d-915d9b0af7ec-kube-api-access-kzk6x\") pod \"redhat-marketplace-px5pg\" (UID: \"aad02c72-91c8-4ce2-b69d-915d9b0af7ec\") " pod="openshift-marketplace/redhat-marketplace-px5pg" Mar 21 05:24:46 crc kubenswrapper[4775]: I0321 05:24:46.555922 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aad02c72-91c8-4ce2-b69d-915d9b0af7ec-catalog-content\") pod \"redhat-marketplace-px5pg\" (UID: \"aad02c72-91c8-4ce2-b69d-915d9b0af7ec\") " pod="openshift-marketplace/redhat-marketplace-px5pg" Mar 21 05:24:46 crc kubenswrapper[4775]: I0321 05:24:46.657804 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aad02c72-91c8-4ce2-b69d-915d9b0af7ec-catalog-content\") pod \"redhat-marketplace-px5pg\" (UID: \"aad02c72-91c8-4ce2-b69d-915d9b0af7ec\") " pod="openshift-marketplace/redhat-marketplace-px5pg" Mar 21 05:24:46 crc kubenswrapper[4775]: I0321 05:24:46.657921 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aad02c72-91c8-4ce2-b69d-915d9b0af7ec-utilities\") pod \"redhat-marketplace-px5pg\" (UID: \"aad02c72-91c8-4ce2-b69d-915d9b0af7ec\") " pod="openshift-marketplace/redhat-marketplace-px5pg" Mar 21 05:24:46 crc kubenswrapper[4775]: I0321 05:24:46.657993 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzk6x\" (UniqueName: \"kubernetes.io/projected/aad02c72-91c8-4ce2-b69d-915d9b0af7ec-kube-api-access-kzk6x\") pod \"redhat-marketplace-px5pg\" (UID: \"aad02c72-91c8-4ce2-b69d-915d9b0af7ec\") " pod="openshift-marketplace/redhat-marketplace-px5pg" Mar 21 05:24:46 crc kubenswrapper[4775]: I0321 05:24:46.658782 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aad02c72-91c8-4ce2-b69d-915d9b0af7ec-catalog-content\") pod \"redhat-marketplace-px5pg\" (UID: \"aad02c72-91c8-4ce2-b69d-915d9b0af7ec\") " pod="openshift-marketplace/redhat-marketplace-px5pg" Mar 21 05:24:46 crc kubenswrapper[4775]: I0321 05:24:46.658831 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aad02c72-91c8-4ce2-b69d-915d9b0af7ec-utilities\") pod \"redhat-marketplace-px5pg\" (UID: \"aad02c72-91c8-4ce2-b69d-915d9b0af7ec\") " pod="openshift-marketplace/redhat-marketplace-px5pg" Mar 21 05:24:46 crc kubenswrapper[4775]: I0321 05:24:46.683013 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzk6x\" (UniqueName: \"kubernetes.io/projected/aad02c72-91c8-4ce2-b69d-915d9b0af7ec-kube-api-access-kzk6x\") pod \"redhat-marketplace-px5pg\" (UID: \"aad02c72-91c8-4ce2-b69d-915d9b0af7ec\") " pod="openshift-marketplace/redhat-marketplace-px5pg" Mar 21 05:24:46 crc kubenswrapper[4775]: I0321 05:24:46.819365 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-px5pg" Mar 21 05:24:47 crc kubenswrapper[4775]: I0321 05:24:47.083526 4775 scope.go:117] "RemoveContainer" containerID="f7a7cb2d632bd295102216dc1df2efdc817134357d27ae93440d6f95f32d6b6b" Mar 21 05:24:47 crc kubenswrapper[4775]: I0321 05:24:47.342497 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-px5pg"] Mar 21 05:24:47 crc kubenswrapper[4775]: I0321 05:24:47.636880 4775 generic.go:334] "Generic (PLEG): container finished" podID="aad02c72-91c8-4ce2-b69d-915d9b0af7ec" containerID="56e35510c790be074d09f5ee8f124d413aaf4e647a0fe13993e83a909111b979" exitCode=0 Mar 21 05:24:47 crc kubenswrapper[4775]: I0321 05:24:47.636930 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-px5pg" event={"ID":"aad02c72-91c8-4ce2-b69d-915d9b0af7ec","Type":"ContainerDied","Data":"56e35510c790be074d09f5ee8f124d413aaf4e647a0fe13993e83a909111b979"} Mar 21 05:24:47 crc kubenswrapper[4775]: I0321 05:24:47.638603 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-px5pg" event={"ID":"aad02c72-91c8-4ce2-b69d-915d9b0af7ec","Type":"ContainerStarted","Data":"a5bd6b6850af17ae87be5c25d646f5af789ea57c2b5ed61ad6e28b7c8a3d545e"} Mar 21 05:24:48 crc kubenswrapper[4775]: I0321 05:24:48.651352 4775 generic.go:334] "Generic (PLEG): container finished" podID="aad02c72-91c8-4ce2-b69d-915d9b0af7ec" containerID="a2a8c8dc1d88f43d6ededf2e8793a512b5087dcbf7e313f245d34485e136258b" exitCode=0 Mar 21 05:24:48 crc kubenswrapper[4775]: I0321 05:24:48.651406 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-px5pg" event={"ID":"aad02c72-91c8-4ce2-b69d-915d9b0af7ec","Type":"ContainerDied","Data":"a2a8c8dc1d88f43d6ededf2e8793a512b5087dcbf7e313f245d34485e136258b"} Mar 21 05:24:49 crc kubenswrapper[4775]: I0321 05:24:49.671618 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-px5pg" event={"ID":"aad02c72-91c8-4ce2-b69d-915d9b0af7ec","Type":"ContainerStarted","Data":"60fb5be3131696012c97db115a74ff472115d80eb7fd9153e9102904c69c91b3"} Mar 21 05:24:49 crc kubenswrapper[4775]: I0321 05:24:49.694875 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-px5pg" podStartSLOduration=2.22205689 podStartE2EDuration="3.694850733s" podCreationTimestamp="2026-03-21 05:24:46 +0000 UTC" firstStartedPulling="2026-03-21 05:24:47.638734106 +0000 UTC m=+2240.615197730" lastFinishedPulling="2026-03-21 05:24:49.111527949 +0000 UTC m=+2242.087991573" observedRunningTime="2026-03-21 05:24:49.682845656 +0000 UTC m=+2242.659309290" watchObservedRunningTime="2026-03-21 05:24:49.694850733 +0000 UTC m=+2242.671314357" Mar 21 05:24:56 crc kubenswrapper[4775]: I0321 05:24:56.820216 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-px5pg" Mar 21 05:24:56 crc kubenswrapper[4775]: I0321 05:24:56.822519 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-px5pg" Mar 21 05:24:56 crc kubenswrapper[4775]: I0321 05:24:56.870223 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-px5pg" Mar 21 05:24:57 crc kubenswrapper[4775]: I0321 05:24:57.793264 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-px5pg" Mar 21 05:24:57 crc kubenswrapper[4775]: I0321 05:24:57.845612 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-px5pg"] Mar 21 05:24:59 crc kubenswrapper[4775]: I0321 05:24:59.752860 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-px5pg" podUID="aad02c72-91c8-4ce2-b69d-915d9b0af7ec" containerName="registry-server" containerID="cri-o://60fb5be3131696012c97db115a74ff472115d80eb7fd9153e9102904c69c91b3" gracePeriod=2 Mar 21 05:25:00 crc kubenswrapper[4775]: I0321 05:25:00.767923 4775 generic.go:334] "Generic (PLEG): container finished" podID="aad02c72-91c8-4ce2-b69d-915d9b0af7ec" containerID="60fb5be3131696012c97db115a74ff472115d80eb7fd9153e9102904c69c91b3" exitCode=0 Mar 21 05:25:00 crc kubenswrapper[4775]: I0321 05:25:00.768173 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-px5pg" event={"ID":"aad02c72-91c8-4ce2-b69d-915d9b0af7ec","Type":"ContainerDied","Data":"60fb5be3131696012c97db115a74ff472115d80eb7fd9153e9102904c69c91b3"} Mar 21 05:25:00 crc kubenswrapper[4775]: I0321 05:25:00.768458 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-px5pg" event={"ID":"aad02c72-91c8-4ce2-b69d-915d9b0af7ec","Type":"ContainerDied","Data":"a5bd6b6850af17ae87be5c25d646f5af789ea57c2b5ed61ad6e28b7c8a3d545e"} Mar 21 05:25:00 crc kubenswrapper[4775]: I0321 05:25:00.768480 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5bd6b6850af17ae87be5c25d646f5af789ea57c2b5ed61ad6e28b7c8a3d545e" Mar 21 05:25:00 crc kubenswrapper[4775]: I0321 05:25:00.783224 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-px5pg" Mar 21 05:25:00 crc kubenswrapper[4775]: I0321 05:25:00.838733 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aad02c72-91c8-4ce2-b69d-915d9b0af7ec-utilities\") pod \"aad02c72-91c8-4ce2-b69d-915d9b0af7ec\" (UID: \"aad02c72-91c8-4ce2-b69d-915d9b0af7ec\") " Mar 21 05:25:00 crc kubenswrapper[4775]: I0321 05:25:00.838792 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aad02c72-91c8-4ce2-b69d-915d9b0af7ec-catalog-content\") pod \"aad02c72-91c8-4ce2-b69d-915d9b0af7ec\" (UID: \"aad02c72-91c8-4ce2-b69d-915d9b0af7ec\") " Mar 21 05:25:00 crc kubenswrapper[4775]: I0321 05:25:00.838843 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzk6x\" (UniqueName: \"kubernetes.io/projected/aad02c72-91c8-4ce2-b69d-915d9b0af7ec-kube-api-access-kzk6x\") pod \"aad02c72-91c8-4ce2-b69d-915d9b0af7ec\" (UID: \"aad02c72-91c8-4ce2-b69d-915d9b0af7ec\") " Mar 21 05:25:00 crc kubenswrapper[4775]: I0321 05:25:00.839752 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aad02c72-91c8-4ce2-b69d-915d9b0af7ec-utilities" (OuterVolumeSpecName: "utilities") pod "aad02c72-91c8-4ce2-b69d-915d9b0af7ec" (UID: "aad02c72-91c8-4ce2-b69d-915d9b0af7ec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:25:00 crc kubenswrapper[4775]: I0321 05:25:00.845828 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aad02c72-91c8-4ce2-b69d-915d9b0af7ec-kube-api-access-kzk6x" (OuterVolumeSpecName: "kube-api-access-kzk6x") pod "aad02c72-91c8-4ce2-b69d-915d9b0af7ec" (UID: "aad02c72-91c8-4ce2-b69d-915d9b0af7ec"). InnerVolumeSpecName "kube-api-access-kzk6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:25:00 crc kubenswrapper[4775]: I0321 05:25:00.869512 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aad02c72-91c8-4ce2-b69d-915d9b0af7ec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aad02c72-91c8-4ce2-b69d-915d9b0af7ec" (UID: "aad02c72-91c8-4ce2-b69d-915d9b0af7ec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:25:00 crc kubenswrapper[4775]: I0321 05:25:00.940914 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aad02c72-91c8-4ce2-b69d-915d9b0af7ec-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:25:00 crc kubenswrapper[4775]: I0321 05:25:00.940985 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aad02c72-91c8-4ce2-b69d-915d9b0af7ec-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:25:00 crc kubenswrapper[4775]: I0321 05:25:00.941008 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzk6x\" (UniqueName: \"kubernetes.io/projected/aad02c72-91c8-4ce2-b69d-915d9b0af7ec-kube-api-access-kzk6x\") on node \"crc\" DevicePath \"\"" Mar 21 05:25:01 crc kubenswrapper[4775]: I0321 05:25:01.777228 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-px5pg" Mar 21 05:25:01 crc kubenswrapper[4775]: I0321 05:25:01.809603 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-px5pg"] Mar 21 05:25:01 crc kubenswrapper[4775]: I0321 05:25:01.819780 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-px5pg"] Mar 21 05:25:02 crc kubenswrapper[4775]: I0321 05:25:02.483372 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:25:02 crc kubenswrapper[4775]: I0321 05:25:02.483474 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:25:03 crc kubenswrapper[4775]: I0321 05:25:03.675104 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aad02c72-91c8-4ce2-b69d-915d9b0af7ec" path="/var/lib/kubelet/pods/aad02c72-91c8-4ce2-b69d-915d9b0af7ec/volumes" Mar 21 05:25:25 crc kubenswrapper[4775]: I0321 05:25:25.763051 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mvbdf"] Mar 21 05:25:25 crc kubenswrapper[4775]: E0321 05:25:25.764266 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aad02c72-91c8-4ce2-b69d-915d9b0af7ec" containerName="extract-utilities" Mar 21 05:25:25 crc kubenswrapper[4775]: I0321 05:25:25.764289 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="aad02c72-91c8-4ce2-b69d-915d9b0af7ec" containerName="extract-utilities" Mar 21 05:25:25 crc kubenswrapper[4775]: E0321 05:25:25.764303 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aad02c72-91c8-4ce2-b69d-915d9b0af7ec" containerName="registry-server" Mar 21 05:25:25 crc kubenswrapper[4775]: I0321 05:25:25.764311 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="aad02c72-91c8-4ce2-b69d-915d9b0af7ec" containerName="registry-server" Mar 21 05:25:25 crc kubenswrapper[4775]: E0321 05:25:25.764343 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aad02c72-91c8-4ce2-b69d-915d9b0af7ec" containerName="extract-content" Mar 21 05:25:25 crc kubenswrapper[4775]: I0321 05:25:25.764353 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="aad02c72-91c8-4ce2-b69d-915d9b0af7ec" containerName="extract-content" Mar 21 05:25:25 crc kubenswrapper[4775]: I0321 05:25:25.764596 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="aad02c72-91c8-4ce2-b69d-915d9b0af7ec" containerName="registry-server" Mar 21 05:25:25 crc kubenswrapper[4775]: I0321 05:25:25.766305 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mvbdf" Mar 21 05:25:25 crc kubenswrapper[4775]: I0321 05:25:25.776391 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mvbdf"] Mar 21 05:25:25 crc kubenswrapper[4775]: I0321 05:25:25.903950 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqrcn\" (UniqueName: \"kubernetes.io/projected/bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f-kube-api-access-dqrcn\") pod \"community-operators-mvbdf\" (UID: \"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f\") " pod="openshift-marketplace/community-operators-mvbdf" Mar 21 05:25:25 crc kubenswrapper[4775]: I0321 05:25:25.904018 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f-utilities\") pod \"community-operators-mvbdf\" (UID: \"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f\") " pod="openshift-marketplace/community-operators-mvbdf" Mar 21 05:25:25 crc kubenswrapper[4775]: I0321 05:25:25.904145 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f-catalog-content\") pod \"community-operators-mvbdf\" (UID: \"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f\") " pod="openshift-marketplace/community-operators-mvbdf" Mar 21 05:25:26 crc kubenswrapper[4775]: I0321 05:25:26.006194 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f-catalog-content\") pod \"community-operators-mvbdf\" (UID: \"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f\") " pod="openshift-marketplace/community-operators-mvbdf" Mar 21 05:25:26 crc kubenswrapper[4775]: I0321 05:25:26.006535 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqrcn\" (UniqueName: \"kubernetes.io/projected/bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f-kube-api-access-dqrcn\") pod \"community-operators-mvbdf\" (UID: \"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f\") " pod="openshift-marketplace/community-operators-mvbdf" Mar 21 05:25:26 crc kubenswrapper[4775]: I0321 05:25:26.006587 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f-utilities\") pod \"community-operators-mvbdf\" (UID: \"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f\") " pod="openshift-marketplace/community-operators-mvbdf" Mar 21 05:25:26 crc kubenswrapper[4775]: I0321 05:25:26.007302 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f-catalog-content\") pod \"community-operators-mvbdf\" (UID: \"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f\") " pod="openshift-marketplace/community-operators-mvbdf" Mar 21 05:25:26 crc kubenswrapper[4775]: I0321 05:25:26.007668 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f-utilities\") pod \"community-operators-mvbdf\" (UID: \"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f\") " pod="openshift-marketplace/community-operators-mvbdf" Mar 21 05:25:26 crc kubenswrapper[4775]: I0321 05:25:26.036439 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqrcn\" (UniqueName: \"kubernetes.io/projected/bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f-kube-api-access-dqrcn\") pod \"community-operators-mvbdf\" (UID: \"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f\") " pod="openshift-marketplace/community-operators-mvbdf" Mar 21 05:25:26 crc kubenswrapper[4775]: I0321 05:25:26.090646 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mvbdf" Mar 21 05:25:26 crc kubenswrapper[4775]: I0321 05:25:26.708196 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mvbdf"] Mar 21 05:25:27 crc kubenswrapper[4775]: I0321 05:25:27.050157 4775 generic.go:334] "Generic (PLEG): container finished" podID="bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f" containerID="f726a1456ad825ba50fc40daddce363b0cc5faa4387d4487cfa9ce395caedbad" exitCode=0 Mar 21 05:25:27 crc kubenswrapper[4775]: I0321 05:25:27.050204 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvbdf" event={"ID":"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f","Type":"ContainerDied","Data":"f726a1456ad825ba50fc40daddce363b0cc5faa4387d4487cfa9ce395caedbad"} Mar 21 05:25:27 crc kubenswrapper[4775]: I0321 05:25:27.050238 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvbdf" event={"ID":"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f","Type":"ContainerStarted","Data":"f2d2898b6bef0be3d1744eb1c72891e1071707d2629275e9a2f5d8e8d9a18417"} Mar 21 05:25:27 crc kubenswrapper[4775]: I0321 05:25:27.052720 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 05:25:28 crc kubenswrapper[4775]: I0321 05:25:28.063590 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvbdf" event={"ID":"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f","Type":"ContainerStarted","Data":"755d96b32d4b2250d80500cc147380b30160a4c2eb540c7f8ac9ac6c15730b9d"} Mar 21 05:25:29 crc kubenswrapper[4775]: I0321 05:25:29.077567 4775 generic.go:334] "Generic (PLEG): container finished" podID="bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f" containerID="755d96b32d4b2250d80500cc147380b30160a4c2eb540c7f8ac9ac6c15730b9d" exitCode=0 Mar 21 05:25:29 crc kubenswrapper[4775]: I0321 05:25:29.077655 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvbdf" event={"ID":"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f","Type":"ContainerDied","Data":"755d96b32d4b2250d80500cc147380b30160a4c2eb540c7f8ac9ac6c15730b9d"} Mar 21 05:25:30 crc kubenswrapper[4775]: I0321 05:25:30.093078 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvbdf" event={"ID":"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f","Type":"ContainerStarted","Data":"69f21971bb6bff290af75da43f3b23e0347527e53e0d039c662e31e7666ba6a6"} Mar 21 05:25:30 crc kubenswrapper[4775]: I0321 05:25:30.120840 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mvbdf" podStartSLOduration=2.686509423 podStartE2EDuration="5.120815107s" podCreationTimestamp="2026-03-21 05:25:25 +0000 UTC" firstStartedPulling="2026-03-21 05:25:27.05236743 +0000 UTC m=+2280.028831054" lastFinishedPulling="2026-03-21 05:25:29.486673114 +0000 UTC m=+2282.463136738" observedRunningTime="2026-03-21 05:25:30.114764407 +0000 UTC m=+2283.091228051" watchObservedRunningTime="2026-03-21 05:25:30.120815107 +0000 UTC m=+2283.097278731" Mar 21 05:25:32 crc kubenswrapper[4775]: I0321 05:25:32.482313 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:25:32 crc kubenswrapper[4775]: I0321 05:25:32.482397 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:25:32 crc kubenswrapper[4775]: I0321 05:25:32.482452 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 05:25:32 crc kubenswrapper[4775]: I0321 05:25:32.483416 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804"} pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 05:25:32 crc kubenswrapper[4775]: I0321 05:25:32.483488 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" containerID="cri-o://21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" gracePeriod=600 Mar 21 05:25:32 crc kubenswrapper[4775]: E0321 05:25:32.625732 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:25:33 crc kubenswrapper[4775]: I0321 05:25:33.123786 4775 generic.go:334] "Generic (PLEG): container finished" podID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" exitCode=0 Mar 21 05:25:33 crc kubenswrapper[4775]: I0321 05:25:33.123848 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerDied","Data":"21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804"} Mar 21 05:25:33 crc kubenswrapper[4775]: I0321 05:25:33.123899 4775 scope.go:117] "RemoveContainer" containerID="820b5e7971de8f534813cc1fdb277aabd37bf9063cedc671ef2c92d0328150cb" Mar 21 05:25:33 crc kubenswrapper[4775]: I0321 05:25:33.124814 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:25:33 crc kubenswrapper[4775]: E0321 05:25:33.125334 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:25:36 crc kubenswrapper[4775]: I0321 05:25:36.091183 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mvbdf" Mar 21 05:25:36 crc kubenswrapper[4775]: I0321 05:25:36.091879 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mvbdf" Mar 21 05:25:36 crc kubenswrapper[4775]: I0321 05:25:36.144241 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mvbdf" Mar 21 05:25:36 crc kubenswrapper[4775]: I0321 05:25:36.226694 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mvbdf" Mar 21 05:25:36 crc kubenswrapper[4775]: I0321 05:25:36.401601 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mvbdf"] Mar 21 05:25:38 crc kubenswrapper[4775]: I0321 05:25:38.177801 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mvbdf" podUID="bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f" containerName="registry-server" containerID="cri-o://69f21971bb6bff290af75da43f3b23e0347527e53e0d039c662e31e7666ba6a6" gracePeriod=2 Mar 21 05:25:38 crc kubenswrapper[4775]: I0321 05:25:38.669059 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mvbdf" Mar 21 05:25:38 crc kubenswrapper[4775]: I0321 05:25:38.777034 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f-catalog-content\") pod \"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f\" (UID: \"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f\") " Mar 21 05:25:38 crc kubenswrapper[4775]: I0321 05:25:38.777101 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f-utilities\") pod \"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f\" (UID: \"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f\") " Mar 21 05:25:38 crc kubenswrapper[4775]: I0321 05:25:38.777286 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqrcn\" (UniqueName: \"kubernetes.io/projected/bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f-kube-api-access-dqrcn\") pod \"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f\" (UID: \"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f\") " Mar 21 05:25:38 crc kubenswrapper[4775]: I0321 05:25:38.778631 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f-utilities" (OuterVolumeSpecName: "utilities") pod "bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f" (UID: "bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:25:38 crc kubenswrapper[4775]: I0321 05:25:38.779487 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:25:38 crc kubenswrapper[4775]: I0321 05:25:38.784506 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f-kube-api-access-dqrcn" (OuterVolumeSpecName: "kube-api-access-dqrcn") pod "bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f" (UID: "bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f"). InnerVolumeSpecName "kube-api-access-dqrcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:25:38 crc kubenswrapper[4775]: I0321 05:25:38.881339 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqrcn\" (UniqueName: \"kubernetes.io/projected/bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f-kube-api-access-dqrcn\") on node \"crc\" DevicePath \"\"" Mar 21 05:25:39 crc kubenswrapper[4775]: I0321 05:25:39.192925 4775 generic.go:334] "Generic (PLEG): container finished" podID="bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f" containerID="69f21971bb6bff290af75da43f3b23e0347527e53e0d039c662e31e7666ba6a6" exitCode=0 Mar 21 05:25:39 crc kubenswrapper[4775]: I0321 05:25:39.192979 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mvbdf" Mar 21 05:25:39 crc kubenswrapper[4775]: I0321 05:25:39.194214 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvbdf" event={"ID":"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f","Type":"ContainerDied","Data":"69f21971bb6bff290af75da43f3b23e0347527e53e0d039c662e31e7666ba6a6"} Mar 21 05:25:39 crc kubenswrapper[4775]: I0321 05:25:39.194328 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvbdf" event={"ID":"bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f","Type":"ContainerDied","Data":"f2d2898b6bef0be3d1744eb1c72891e1071707d2629275e9a2f5d8e8d9a18417"} Mar 21 05:25:39 crc kubenswrapper[4775]: I0321 05:25:39.194414 4775 scope.go:117] "RemoveContainer" containerID="69f21971bb6bff290af75da43f3b23e0347527e53e0d039c662e31e7666ba6a6" Mar 21 05:25:39 crc kubenswrapper[4775]: I0321 05:25:39.217541 4775 scope.go:117] "RemoveContainer" containerID="755d96b32d4b2250d80500cc147380b30160a4c2eb540c7f8ac9ac6c15730b9d" Mar 21 05:25:39 crc kubenswrapper[4775]: I0321 05:25:39.241962 4775 scope.go:117] "RemoveContainer" containerID="f726a1456ad825ba50fc40daddce363b0cc5faa4387d4487cfa9ce395caedbad" Mar 21 05:25:39 crc kubenswrapper[4775]: I0321 05:25:39.299297 4775 scope.go:117] "RemoveContainer" containerID="69f21971bb6bff290af75da43f3b23e0347527e53e0d039c662e31e7666ba6a6" Mar 21 05:25:39 crc kubenswrapper[4775]: E0321 05:25:39.299890 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69f21971bb6bff290af75da43f3b23e0347527e53e0d039c662e31e7666ba6a6\": container with ID starting with 69f21971bb6bff290af75da43f3b23e0347527e53e0d039c662e31e7666ba6a6 not found: ID does not exist" containerID="69f21971bb6bff290af75da43f3b23e0347527e53e0d039c662e31e7666ba6a6" Mar 21 05:25:39 crc kubenswrapper[4775]: I0321 05:25:39.299945 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69f21971bb6bff290af75da43f3b23e0347527e53e0d039c662e31e7666ba6a6"} err="failed to get container status \"69f21971bb6bff290af75da43f3b23e0347527e53e0d039c662e31e7666ba6a6\": rpc error: code = NotFound desc = could not find container \"69f21971bb6bff290af75da43f3b23e0347527e53e0d039c662e31e7666ba6a6\": container with ID starting with 69f21971bb6bff290af75da43f3b23e0347527e53e0d039c662e31e7666ba6a6 not found: ID does not exist" Mar 21 05:25:39 crc kubenswrapper[4775]: I0321 05:25:39.299978 4775 scope.go:117] "RemoveContainer" containerID="755d96b32d4b2250d80500cc147380b30160a4c2eb540c7f8ac9ac6c15730b9d" Mar 21 05:25:39 crc kubenswrapper[4775]: E0321 05:25:39.300322 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"755d96b32d4b2250d80500cc147380b30160a4c2eb540c7f8ac9ac6c15730b9d\": container with ID starting with 755d96b32d4b2250d80500cc147380b30160a4c2eb540c7f8ac9ac6c15730b9d not found: ID does not exist" containerID="755d96b32d4b2250d80500cc147380b30160a4c2eb540c7f8ac9ac6c15730b9d" Mar 21 05:25:39 crc kubenswrapper[4775]: I0321 05:25:39.300377 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"755d96b32d4b2250d80500cc147380b30160a4c2eb540c7f8ac9ac6c15730b9d"} err="failed to get container status \"755d96b32d4b2250d80500cc147380b30160a4c2eb540c7f8ac9ac6c15730b9d\": rpc error: code = NotFound desc = could not find container \"755d96b32d4b2250d80500cc147380b30160a4c2eb540c7f8ac9ac6c15730b9d\": container with ID starting with 755d96b32d4b2250d80500cc147380b30160a4c2eb540c7f8ac9ac6c15730b9d not found: ID does not exist" Mar 21 05:25:39 crc kubenswrapper[4775]: I0321 05:25:39.300396 4775 scope.go:117] "RemoveContainer" containerID="f726a1456ad825ba50fc40daddce363b0cc5faa4387d4487cfa9ce395caedbad" Mar 21 05:25:39 crc kubenswrapper[4775]: E0321 05:25:39.300631 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f726a1456ad825ba50fc40daddce363b0cc5faa4387d4487cfa9ce395caedbad\": container with ID starting with f726a1456ad825ba50fc40daddce363b0cc5faa4387d4487cfa9ce395caedbad not found: ID does not exist" containerID="f726a1456ad825ba50fc40daddce363b0cc5faa4387d4487cfa9ce395caedbad" Mar 21 05:25:39 crc kubenswrapper[4775]: I0321 05:25:39.300653 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f726a1456ad825ba50fc40daddce363b0cc5faa4387d4487cfa9ce395caedbad"} err="failed to get container status \"f726a1456ad825ba50fc40daddce363b0cc5faa4387d4487cfa9ce395caedbad\": rpc error: code = NotFound desc = could not find container \"f726a1456ad825ba50fc40daddce363b0cc5faa4387d4487cfa9ce395caedbad\": container with ID starting with f726a1456ad825ba50fc40daddce363b0cc5faa4387d4487cfa9ce395caedbad not found: ID does not exist" Mar 21 05:25:39 crc kubenswrapper[4775]: I0321 05:25:39.437220 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f" (UID: "bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:25:39 crc kubenswrapper[4775]: I0321 05:25:39.494704 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:25:39 crc kubenswrapper[4775]: I0321 05:25:39.535184 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mvbdf"] Mar 21 05:25:39 crc kubenswrapper[4775]: I0321 05:25:39.544581 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mvbdf"] Mar 21 05:25:39 crc kubenswrapper[4775]: I0321 05:25:39.675106 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f" path="/var/lib/kubelet/pods/bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f/volumes" Mar 21 05:25:48 crc kubenswrapper[4775]: I0321 05:25:48.662468 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:25:48 crc kubenswrapper[4775]: E0321 05:25:48.663594 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:26:00 crc kubenswrapper[4775]: I0321 05:26:00.155392 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567846-6lslz"] Mar 21 05:26:00 crc kubenswrapper[4775]: E0321 05:26:00.156445 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f" containerName="registry-server" Mar 21 05:26:00 crc kubenswrapper[4775]: I0321 05:26:00.156461 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f" containerName="registry-server" Mar 21 05:26:00 crc kubenswrapper[4775]: E0321 05:26:00.156483 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f" containerName="extract-utilities" Mar 21 05:26:00 crc kubenswrapper[4775]: I0321 05:26:00.156489 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f" containerName="extract-utilities" Mar 21 05:26:00 crc kubenswrapper[4775]: E0321 05:26:00.156508 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f" containerName="extract-content" Mar 21 05:26:00 crc kubenswrapper[4775]: I0321 05:26:00.156516 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f" containerName="extract-content" Mar 21 05:26:00 crc kubenswrapper[4775]: I0321 05:26:00.156722 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad21dd2-8ec4-407c-a1d5-8c617dd4fc1f" containerName="registry-server" Mar 21 05:26:00 crc kubenswrapper[4775]: I0321 05:26:00.157415 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567846-6lslz" Mar 21 05:26:00 crc kubenswrapper[4775]: I0321 05:26:00.163900 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:26:00 crc kubenswrapper[4775]: I0321 05:26:00.164184 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:26:00 crc kubenswrapper[4775]: I0321 05:26:00.164987 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:26:00 crc kubenswrapper[4775]: I0321 05:26:00.165830 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567846-6lslz"] Mar 21 05:26:00 crc kubenswrapper[4775]: I0321 05:26:00.263965 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nr2wl\" (UniqueName: \"kubernetes.io/projected/43f89ba5-2605-4bdb-bc77-020c6bf9db75-kube-api-access-nr2wl\") pod \"auto-csr-approver-29567846-6lslz\" (UID: \"43f89ba5-2605-4bdb-bc77-020c6bf9db75\") " pod="openshift-infra/auto-csr-approver-29567846-6lslz" Mar 21 05:26:00 crc kubenswrapper[4775]: I0321 05:26:00.366308 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nr2wl\" (UniqueName: \"kubernetes.io/projected/43f89ba5-2605-4bdb-bc77-020c6bf9db75-kube-api-access-nr2wl\") pod \"auto-csr-approver-29567846-6lslz\" (UID: \"43f89ba5-2605-4bdb-bc77-020c6bf9db75\") " pod="openshift-infra/auto-csr-approver-29567846-6lslz" Mar 21 05:26:00 crc kubenswrapper[4775]: I0321 05:26:00.389431 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nr2wl\" (UniqueName: \"kubernetes.io/projected/43f89ba5-2605-4bdb-bc77-020c6bf9db75-kube-api-access-nr2wl\") pod \"auto-csr-approver-29567846-6lslz\" (UID: \"43f89ba5-2605-4bdb-bc77-020c6bf9db75\") " pod="openshift-infra/auto-csr-approver-29567846-6lslz" Mar 21 05:26:00 crc kubenswrapper[4775]: I0321 05:26:00.492786 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567846-6lslz" Mar 21 05:26:00 crc kubenswrapper[4775]: I0321 05:26:00.661429 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:26:00 crc kubenswrapper[4775]: E0321 05:26:00.662455 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:26:00 crc kubenswrapper[4775]: I0321 05:26:00.956061 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567846-6lslz"] Mar 21 05:26:01 crc kubenswrapper[4775]: I0321 05:26:01.405339 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567846-6lslz" event={"ID":"43f89ba5-2605-4bdb-bc77-020c6bf9db75","Type":"ContainerStarted","Data":"a0cf3c13c79d3900fe284a3ffd7e6e9a637b796bfedf0ece83845a7f24aeae65"} Mar 21 05:26:02 crc kubenswrapper[4775]: I0321 05:26:02.418045 4775 generic.go:334] "Generic (PLEG): container finished" podID="43f89ba5-2605-4bdb-bc77-020c6bf9db75" containerID="a8931f9c5d17f0e3f84a26748f81ac4ac575c43d3a88da2c5e6613b2fee09eee" exitCode=0 Mar 21 05:26:02 crc kubenswrapper[4775]: I0321 05:26:02.418107 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567846-6lslz" event={"ID":"43f89ba5-2605-4bdb-bc77-020c6bf9db75","Type":"ContainerDied","Data":"a8931f9c5d17f0e3f84a26748f81ac4ac575c43d3a88da2c5e6613b2fee09eee"} Mar 21 05:26:03 crc kubenswrapper[4775]: I0321 05:26:03.808792 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567846-6lslz" Mar 21 05:26:03 crc kubenswrapper[4775]: I0321 05:26:03.942808 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nr2wl\" (UniqueName: \"kubernetes.io/projected/43f89ba5-2605-4bdb-bc77-020c6bf9db75-kube-api-access-nr2wl\") pod \"43f89ba5-2605-4bdb-bc77-020c6bf9db75\" (UID: \"43f89ba5-2605-4bdb-bc77-020c6bf9db75\") " Mar 21 05:26:03 crc kubenswrapper[4775]: I0321 05:26:03.955286 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43f89ba5-2605-4bdb-bc77-020c6bf9db75-kube-api-access-nr2wl" (OuterVolumeSpecName: "kube-api-access-nr2wl") pod "43f89ba5-2605-4bdb-bc77-020c6bf9db75" (UID: "43f89ba5-2605-4bdb-bc77-020c6bf9db75"). InnerVolumeSpecName "kube-api-access-nr2wl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:26:04 crc kubenswrapper[4775]: I0321 05:26:04.046154 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nr2wl\" (UniqueName: \"kubernetes.io/projected/43f89ba5-2605-4bdb-bc77-020c6bf9db75-kube-api-access-nr2wl\") on node \"crc\" DevicePath \"\"" Mar 21 05:26:04 crc kubenswrapper[4775]: I0321 05:26:04.438959 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567846-6lslz" event={"ID":"43f89ba5-2605-4bdb-bc77-020c6bf9db75","Type":"ContainerDied","Data":"a0cf3c13c79d3900fe284a3ffd7e6e9a637b796bfedf0ece83845a7f24aeae65"} Mar 21 05:26:04 crc kubenswrapper[4775]: I0321 05:26:04.439004 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0cf3c13c79d3900fe284a3ffd7e6e9a637b796bfedf0ece83845a7f24aeae65" Mar 21 05:26:04 crc kubenswrapper[4775]: I0321 05:26:04.438999 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567846-6lslz" Mar 21 05:26:04 crc kubenswrapper[4775]: I0321 05:26:04.888225 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567840-lm4bc"] Mar 21 05:26:04 crc kubenswrapper[4775]: I0321 05:26:04.898111 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567840-lm4bc"] Mar 21 05:26:05 crc kubenswrapper[4775]: I0321 05:26:05.673906 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90536c4e-c461-4d2c-95dd-f08660cb2e69" path="/var/lib/kubelet/pods/90536c4e-c461-4d2c-95dd-f08660cb2e69/volumes" Mar 21 05:26:11 crc kubenswrapper[4775]: I0321 05:26:11.662846 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:26:11 crc kubenswrapper[4775]: E0321 05:26:11.664298 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:26:26 crc kubenswrapper[4775]: I0321 05:26:26.661891 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:26:26 crc kubenswrapper[4775]: E0321 05:26:26.662809 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:26:40 crc kubenswrapper[4775]: I0321 05:26:40.662477 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:26:40 crc kubenswrapper[4775]: E0321 05:26:40.663321 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:26:47 crc kubenswrapper[4775]: I0321 05:26:47.239136 4775 scope.go:117] "RemoveContainer" containerID="383196299bf8ab4c199e90d21197235aa0a2587594bef58700d15b41f3c4abec" Mar 21 05:26:53 crc kubenswrapper[4775]: I0321 05:26:53.661721 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:26:53 crc kubenswrapper[4775]: E0321 05:26:53.662755 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:27:04 crc kubenswrapper[4775]: I0321 05:27:04.661788 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:27:04 crc kubenswrapper[4775]: E0321 05:27:04.662879 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:27:15 crc kubenswrapper[4775]: I0321 05:27:15.662876 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:27:15 crc kubenswrapper[4775]: E0321 05:27:15.663743 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:27:25 crc kubenswrapper[4775]: I0321 05:27:25.418913 4775 generic.go:334] "Generic (PLEG): container finished" podID="a71dcc90-c70a-4ff8-bf4a-42f1a2415827" containerID="1e34cd56643c271492ee835c6208defe78742763b335409f4aab2ec44574dc78" exitCode=0 Mar 21 05:27:25 crc kubenswrapper[4775]: I0321 05:27:25.419001 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" event={"ID":"a71dcc90-c70a-4ff8-bf4a-42f1a2415827","Type":"ContainerDied","Data":"1e34cd56643c271492ee835c6208defe78742763b335409f4aab2ec44574dc78"} Mar 21 05:27:26 crc kubenswrapper[4775]: I0321 05:27:26.981695 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.011192 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-libvirt-combined-ca-bundle\") pod \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\" (UID: \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\") " Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.011575 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f844j\" (UniqueName: \"kubernetes.io/projected/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-kube-api-access-f844j\") pod \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\" (UID: \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\") " Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.011655 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-inventory\") pod \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\" (UID: \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\") " Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.011814 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-ssh-key-openstack-edpm-ipam\") pod \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\" (UID: \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\") " Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.011885 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-libvirt-secret-0\") pod \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\" (UID: \"a71dcc90-c70a-4ff8-bf4a-42f1a2415827\") " Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.019544 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-kube-api-access-f844j" (OuterVolumeSpecName: "kube-api-access-f844j") pod "a71dcc90-c70a-4ff8-bf4a-42f1a2415827" (UID: "a71dcc90-c70a-4ff8-bf4a-42f1a2415827"). InnerVolumeSpecName "kube-api-access-f844j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.026502 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "a71dcc90-c70a-4ff8-bf4a-42f1a2415827" (UID: "a71dcc90-c70a-4ff8-bf4a-42f1a2415827"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.049405 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-inventory" (OuterVolumeSpecName: "inventory") pod "a71dcc90-c70a-4ff8-bf4a-42f1a2415827" (UID: "a71dcc90-c70a-4ff8-bf4a-42f1a2415827"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.058971 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "a71dcc90-c70a-4ff8-bf4a-42f1a2415827" (UID: "a71dcc90-c70a-4ff8-bf4a-42f1a2415827"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.069584 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a71dcc90-c70a-4ff8-bf4a-42f1a2415827" (UID: "a71dcc90-c70a-4ff8-bf4a-42f1a2415827"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.121157 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.121208 4775 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.121217 4775 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.121231 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f844j\" (UniqueName: \"kubernetes.io/projected/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-kube-api-access-f844j\") on node \"crc\" DevicePath \"\"" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.121242 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a71dcc90-c70a-4ff8-bf4a-42f1a2415827-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.442193 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" event={"ID":"a71dcc90-c70a-4ff8-bf4a-42f1a2415827","Type":"ContainerDied","Data":"7a5a248239d97b47a9a8821636d4ceb0030545a6d3ce4d3885e064cebb074b6b"} Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.442260 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a5a248239d97b47a9a8821636d4ceb0030545a6d3ce4d3885e064cebb074b6b" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.442368 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.565627 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7"] Mar 21 05:27:27 crc kubenswrapper[4775]: E0321 05:27:27.566367 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a71dcc90-c70a-4ff8-bf4a-42f1a2415827" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.566390 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a71dcc90-c70a-4ff8-bf4a-42f1a2415827" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 21 05:27:27 crc kubenswrapper[4775]: E0321 05:27:27.566408 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43f89ba5-2605-4bdb-bc77-020c6bf9db75" containerName="oc" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.566415 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="43f89ba5-2605-4bdb-bc77-020c6bf9db75" containerName="oc" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.566665 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="43f89ba5-2605-4bdb-bc77-020c6bf9db75" containerName="oc" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.566683 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a71dcc90-c70a-4ff8-bf4a-42f1a2415827" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.567718 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.570520 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.570656 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.570764 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.571025 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.571274 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljtt5" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.573312 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.576639 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.586885 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7"] Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.632754 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.632823 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.632883 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/50003f97-774c-4321-9ddf-6ac67546b19f-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.632909 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.632938 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.632967 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9d49\" (UniqueName: \"kubernetes.io/projected/50003f97-774c-4321-9ddf-6ac67546b19f-kube-api-access-t9d49\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.632991 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.633024 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.633058 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.633085 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.633106 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.735820 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.735912 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.735990 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.736239 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.736321 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.736396 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/50003f97-774c-4321-9ddf-6ac67546b19f-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.736425 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.736472 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.736510 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9d49\" (UniqueName: \"kubernetes.io/projected/50003f97-774c-4321-9ddf-6ac67546b19f-kube-api-access-t9d49\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.736544 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.736586 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.742853 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.742908 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.742967 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.743011 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.743159 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.745191 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.754549 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.754721 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.755650 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/50003f97-774c-4321-9ddf-6ac67546b19f-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.756030 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.756547 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.757374 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.761768 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.762197 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.762719 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9d49\" (UniqueName: \"kubernetes.io/projected/50003f97-774c-4321-9ddf-6ac67546b19f-kube-api-access-t9d49\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.765753 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-t85b7\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.889316 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljtt5" Mar 21 05:27:27 crc kubenswrapper[4775]: I0321 05:27:27.896691 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:27:28 crc kubenswrapper[4775]: I0321 05:27:28.463770 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7"] Mar 21 05:27:28 crc kubenswrapper[4775]: I0321 05:27:28.956869 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 05:27:29 crc kubenswrapper[4775]: I0321 05:27:29.465394 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" event={"ID":"50003f97-774c-4321-9ddf-6ac67546b19f","Type":"ContainerStarted","Data":"1bd31fe06833c1b1b26bf4c74b0a110b0c9285cb71f3d43fb71c6f929a94bde5"} Mar 21 05:27:29 crc kubenswrapper[4775]: I0321 05:27:29.465750 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" event={"ID":"50003f97-774c-4321-9ddf-6ac67546b19f","Type":"ContainerStarted","Data":"05426bdb669ddf8a7c54efd23e97575c7cd08628924364b8b1f0fca7b46f8ad1"} Mar 21 05:27:29 crc kubenswrapper[4775]: I0321 05:27:29.488367 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" podStartSLOduration=2.00543805 podStartE2EDuration="2.488344856s" podCreationTimestamp="2026-03-21 05:27:27 +0000 UTC" firstStartedPulling="2026-03-21 05:27:28.47030355 +0000 UTC m=+2401.446767174" lastFinishedPulling="2026-03-21 05:27:28.953210356 +0000 UTC m=+2401.929673980" observedRunningTime="2026-03-21 05:27:29.48241454 +0000 UTC m=+2402.458878164" watchObservedRunningTime="2026-03-21 05:27:29.488344856 +0000 UTC m=+2402.464808480" Mar 21 05:27:30 crc kubenswrapper[4775]: I0321 05:27:30.662057 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:27:30 crc kubenswrapper[4775]: E0321 05:27:30.662793 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:27:43 crc kubenswrapper[4775]: I0321 05:27:43.661724 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:27:43 crc kubenswrapper[4775]: E0321 05:27:43.662677 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:27:54 crc kubenswrapper[4775]: I0321 05:27:54.662247 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:27:54 crc kubenswrapper[4775]: E0321 05:27:54.663301 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:28:00 crc kubenswrapper[4775]: I0321 05:28:00.160289 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567848-6lggf"] Mar 21 05:28:00 crc kubenswrapper[4775]: I0321 05:28:00.162436 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567848-6lggf" Mar 21 05:28:00 crc kubenswrapper[4775]: I0321 05:28:00.165813 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:28:00 crc kubenswrapper[4775]: I0321 05:28:00.167526 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:28:00 crc kubenswrapper[4775]: I0321 05:28:00.168020 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:28:00 crc kubenswrapper[4775]: I0321 05:28:00.173555 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567848-6lggf"] Mar 21 05:28:00 crc kubenswrapper[4775]: I0321 05:28:00.294092 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6wwl\" (UniqueName: \"kubernetes.io/projected/5d6c8144-5186-4a29-a208-b929f47b1695-kube-api-access-j6wwl\") pod \"auto-csr-approver-29567848-6lggf\" (UID: \"5d6c8144-5186-4a29-a208-b929f47b1695\") " pod="openshift-infra/auto-csr-approver-29567848-6lggf" Mar 21 05:28:00 crc kubenswrapper[4775]: I0321 05:28:00.397721 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6wwl\" (UniqueName: \"kubernetes.io/projected/5d6c8144-5186-4a29-a208-b929f47b1695-kube-api-access-j6wwl\") pod \"auto-csr-approver-29567848-6lggf\" (UID: \"5d6c8144-5186-4a29-a208-b929f47b1695\") " pod="openshift-infra/auto-csr-approver-29567848-6lggf" Mar 21 05:28:00 crc kubenswrapper[4775]: I0321 05:28:00.426094 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6wwl\" (UniqueName: \"kubernetes.io/projected/5d6c8144-5186-4a29-a208-b929f47b1695-kube-api-access-j6wwl\") pod \"auto-csr-approver-29567848-6lggf\" (UID: \"5d6c8144-5186-4a29-a208-b929f47b1695\") " pod="openshift-infra/auto-csr-approver-29567848-6lggf" Mar 21 05:28:00 crc kubenswrapper[4775]: I0321 05:28:00.494017 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567848-6lggf" Mar 21 05:28:00 crc kubenswrapper[4775]: I0321 05:28:00.962802 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567848-6lggf"] Mar 21 05:28:01 crc kubenswrapper[4775]: I0321 05:28:01.803292 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567848-6lggf" event={"ID":"5d6c8144-5186-4a29-a208-b929f47b1695","Type":"ContainerStarted","Data":"8461d8c1fa111103e9f9eae3a8d53c59bf4bf66b5c81f1cafcc6ae54ee1679aa"} Mar 21 05:28:02 crc kubenswrapper[4775]: I0321 05:28:02.827518 4775 generic.go:334] "Generic (PLEG): container finished" podID="5d6c8144-5186-4a29-a208-b929f47b1695" containerID="b0bde1a2f3bc4718dd2d5ab50594a32268b2724339820463f90a790bd3794b31" exitCode=0 Mar 21 05:28:02 crc kubenswrapper[4775]: I0321 05:28:02.828219 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567848-6lggf" event={"ID":"5d6c8144-5186-4a29-a208-b929f47b1695","Type":"ContainerDied","Data":"b0bde1a2f3bc4718dd2d5ab50594a32268b2724339820463f90a790bd3794b31"} Mar 21 05:28:04 crc kubenswrapper[4775]: I0321 05:28:04.147913 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567848-6lggf" Mar 21 05:28:04 crc kubenswrapper[4775]: I0321 05:28:04.291887 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6wwl\" (UniqueName: \"kubernetes.io/projected/5d6c8144-5186-4a29-a208-b929f47b1695-kube-api-access-j6wwl\") pod \"5d6c8144-5186-4a29-a208-b929f47b1695\" (UID: \"5d6c8144-5186-4a29-a208-b929f47b1695\") " Mar 21 05:28:04 crc kubenswrapper[4775]: I0321 05:28:04.303334 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d6c8144-5186-4a29-a208-b929f47b1695-kube-api-access-j6wwl" (OuterVolumeSpecName: "kube-api-access-j6wwl") pod "5d6c8144-5186-4a29-a208-b929f47b1695" (UID: "5d6c8144-5186-4a29-a208-b929f47b1695"). InnerVolumeSpecName "kube-api-access-j6wwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:28:04 crc kubenswrapper[4775]: I0321 05:28:04.395711 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6wwl\" (UniqueName: \"kubernetes.io/projected/5d6c8144-5186-4a29-a208-b929f47b1695-kube-api-access-j6wwl\") on node \"crc\" DevicePath \"\"" Mar 21 05:28:04 crc kubenswrapper[4775]: I0321 05:28:04.851249 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567848-6lggf" event={"ID":"5d6c8144-5186-4a29-a208-b929f47b1695","Type":"ContainerDied","Data":"8461d8c1fa111103e9f9eae3a8d53c59bf4bf66b5c81f1cafcc6ae54ee1679aa"} Mar 21 05:28:04 crc kubenswrapper[4775]: I0321 05:28:04.851297 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8461d8c1fa111103e9f9eae3a8d53c59bf4bf66b5c81f1cafcc6ae54ee1679aa" Mar 21 05:28:04 crc kubenswrapper[4775]: I0321 05:28:04.851355 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567848-6lggf" Mar 21 05:28:05 crc kubenswrapper[4775]: I0321 05:28:05.228967 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567842-zs7vw"] Mar 21 05:28:05 crc kubenswrapper[4775]: I0321 05:28:05.241197 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567842-zs7vw"] Mar 21 05:28:05 crc kubenswrapper[4775]: I0321 05:28:05.672456 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="616ac3d7-a66e-45e3-b9ca-257dbd29e212" path="/var/lib/kubelet/pods/616ac3d7-a66e-45e3-b9ca-257dbd29e212/volumes" Mar 21 05:28:07 crc kubenswrapper[4775]: I0321 05:28:07.670177 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:28:07 crc kubenswrapper[4775]: E0321 05:28:07.670661 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:28:20 crc kubenswrapper[4775]: I0321 05:28:20.662142 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:28:20 crc kubenswrapper[4775]: E0321 05:28:20.662876 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:28:35 crc kubenswrapper[4775]: I0321 05:28:35.662632 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:28:35 crc kubenswrapper[4775]: E0321 05:28:35.664406 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:28:47 crc kubenswrapper[4775]: I0321 05:28:47.350211 4775 scope.go:117] "RemoveContainer" containerID="a383e6d9b0bff1f3c71589e15f0f0c78443f3a1f70c9b149974bd8327eb1f252" Mar 21 05:28:50 crc kubenswrapper[4775]: I0321 05:28:50.662485 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:28:50 crc kubenswrapper[4775]: E0321 05:28:50.663338 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:29:02 crc kubenswrapper[4775]: I0321 05:29:02.661947 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:29:02 crc kubenswrapper[4775]: E0321 05:29:02.663111 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:29:16 crc kubenswrapper[4775]: I0321 05:29:16.662090 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:29:16 crc kubenswrapper[4775]: E0321 05:29:16.663308 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:29:31 crc kubenswrapper[4775]: I0321 05:29:31.661917 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:29:31 crc kubenswrapper[4775]: E0321 05:29:31.662918 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:29:42 crc kubenswrapper[4775]: I0321 05:29:42.662973 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:29:42 crc kubenswrapper[4775]: E0321 05:29:42.663960 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:29:45 crc kubenswrapper[4775]: I0321 05:29:45.133823 4775 generic.go:334] "Generic (PLEG): container finished" podID="50003f97-774c-4321-9ddf-6ac67546b19f" containerID="1bd31fe06833c1b1b26bf4c74b0a110b0c9285cb71f3d43fb71c6f929a94bde5" exitCode=0 Mar 21 05:29:45 crc kubenswrapper[4775]: I0321 05:29:45.134319 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" event={"ID":"50003f97-774c-4321-9ddf-6ac67546b19f","Type":"ContainerDied","Data":"1bd31fe06833c1b1b26bf4c74b0a110b0c9285cb71f3d43fb71c6f929a94bde5"} Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.632371 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.732442 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-0\") pod \"50003f97-774c-4321-9ddf-6ac67546b19f\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.732522 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9d49\" (UniqueName: \"kubernetes.io/projected/50003f97-774c-4321-9ddf-6ac67546b19f-kube-api-access-t9d49\") pod \"50003f97-774c-4321-9ddf-6ac67546b19f\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.732573 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-ssh-key-openstack-edpm-ipam\") pod \"50003f97-774c-4321-9ddf-6ac67546b19f\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.732612 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-1\") pod \"50003f97-774c-4321-9ddf-6ac67546b19f\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.732635 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-2\") pod \"50003f97-774c-4321-9ddf-6ac67546b19f\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.732697 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-combined-ca-bundle\") pod \"50003f97-774c-4321-9ddf-6ac67546b19f\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.732721 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/50003f97-774c-4321-9ddf-6ac67546b19f-nova-extra-config-0\") pod \"50003f97-774c-4321-9ddf-6ac67546b19f\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.732752 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-3\") pod \"50003f97-774c-4321-9ddf-6ac67546b19f\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.732858 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-migration-ssh-key-1\") pod \"50003f97-774c-4321-9ddf-6ac67546b19f\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.733001 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-migration-ssh-key-0\") pod \"50003f97-774c-4321-9ddf-6ac67546b19f\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.733056 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-inventory\") pod \"50003f97-774c-4321-9ddf-6ac67546b19f\" (UID: \"50003f97-774c-4321-9ddf-6ac67546b19f\") " Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.750704 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50003f97-774c-4321-9ddf-6ac67546b19f-kube-api-access-t9d49" (OuterVolumeSpecName: "kube-api-access-t9d49") pod "50003f97-774c-4321-9ddf-6ac67546b19f" (UID: "50003f97-774c-4321-9ddf-6ac67546b19f"). InnerVolumeSpecName "kube-api-access-t9d49". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.756319 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "50003f97-774c-4321-9ddf-6ac67546b19f" (UID: "50003f97-774c-4321-9ddf-6ac67546b19f"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.766988 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "50003f97-774c-4321-9ddf-6ac67546b19f" (UID: "50003f97-774c-4321-9ddf-6ac67546b19f"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.770872 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "50003f97-774c-4321-9ddf-6ac67546b19f" (UID: "50003f97-774c-4321-9ddf-6ac67546b19f"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.777497 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "50003f97-774c-4321-9ddf-6ac67546b19f" (UID: "50003f97-774c-4321-9ddf-6ac67546b19f"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.779829 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "50003f97-774c-4321-9ddf-6ac67546b19f" (UID: "50003f97-774c-4321-9ddf-6ac67546b19f"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.782434 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-inventory" (OuterVolumeSpecName: "inventory") pod "50003f97-774c-4321-9ddf-6ac67546b19f" (UID: "50003f97-774c-4321-9ddf-6ac67546b19f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.782656 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "50003f97-774c-4321-9ddf-6ac67546b19f" (UID: "50003f97-774c-4321-9ddf-6ac67546b19f"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.792756 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50003f97-774c-4321-9ddf-6ac67546b19f-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "50003f97-774c-4321-9ddf-6ac67546b19f" (UID: "50003f97-774c-4321-9ddf-6ac67546b19f"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.804032 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "50003f97-774c-4321-9ddf-6ac67546b19f" (UID: "50003f97-774c-4321-9ddf-6ac67546b19f"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.805994 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "50003f97-774c-4321-9ddf-6ac67546b19f" (UID: "50003f97-774c-4321-9ddf-6ac67546b19f"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.835705 4775 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.835756 4775 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.835771 4775 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.835780 4775 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/50003f97-774c-4321-9ddf-6ac67546b19f-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.835794 4775 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.835804 4775 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.835813 4775 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.835825 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.835837 4775 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.835848 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9d49\" (UniqueName: \"kubernetes.io/projected/50003f97-774c-4321-9ddf-6ac67546b19f-kube-api-access-t9d49\") on node \"crc\" DevicePath \"\"" Mar 21 05:29:46 crc kubenswrapper[4775]: I0321 05:29:46.835860 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50003f97-774c-4321-9ddf-6ac67546b19f-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.153174 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" event={"ID":"50003f97-774c-4321-9ddf-6ac67546b19f","Type":"ContainerDied","Data":"05426bdb669ddf8a7c54efd23e97575c7cd08628924364b8b1f0fca7b46f8ad1"} Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.153222 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05426bdb669ddf8a7c54efd23e97575c7cd08628924364b8b1f0fca7b46f8ad1" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.153221 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-t85b7" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.286537 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg"] Mar 21 05:29:47 crc kubenswrapper[4775]: E0321 05:29:47.286979 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d6c8144-5186-4a29-a208-b929f47b1695" containerName="oc" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.286998 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d6c8144-5186-4a29-a208-b929f47b1695" containerName="oc" Mar 21 05:29:47 crc kubenswrapper[4775]: E0321 05:29:47.287040 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50003f97-774c-4321-9ddf-6ac67546b19f" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.287049 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="50003f97-774c-4321-9ddf-6ac67546b19f" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.287411 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="50003f97-774c-4321-9ddf-6ac67546b19f" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.287435 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d6c8144-5186-4a29-a208-b929f47b1695" containerName="oc" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.288559 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.292994 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.293098 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.293016 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.293342 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.293555 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ljtt5" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.316307 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg"] Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.348597 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.348650 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.348697 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.348932 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pskfk\" (UniqueName: \"kubernetes.io/projected/bccbefa9-966d-44b9-bd8f-bb566649b315-kube-api-access-pskfk\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.349113 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.349374 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.349642 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.432518 4775 scope.go:117] "RemoveContainer" containerID="441225cb122bf66cabf5c48f54264da9293c307b82c5b70551d4830941ce387b" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.452601 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.452720 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pskfk\" (UniqueName: \"kubernetes.io/projected/bccbefa9-966d-44b9-bd8f-bb566649b315-kube-api-access-pskfk\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.452789 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.452902 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.453042 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.453157 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.453202 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.460827 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.462608 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.463223 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.464846 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.465196 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.469194 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.485202 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pskfk\" (UniqueName: \"kubernetes.io/projected/bccbefa9-966d-44b9-bd8f-bb566649b315-kube-api-access-pskfk\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.486453 4775 scope.go:117] "RemoveContainer" containerID="a85357de462c6ddc32de73140ccff74411e8b1a6bb88a57b428ef54e3561339b" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.518805 4775 scope.go:117] "RemoveContainer" containerID="cd99f6278b5762bb26ce5bb130e6b5017494ab8ebb62ead08c30e6235d692bd9" Mar 21 05:29:47 crc kubenswrapper[4775]: I0321 05:29:47.615024 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:29:48 crc kubenswrapper[4775]: I0321 05:29:48.179704 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg"] Mar 21 05:29:49 crc kubenswrapper[4775]: I0321 05:29:49.174304 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" event={"ID":"bccbefa9-966d-44b9-bd8f-bb566649b315","Type":"ContainerStarted","Data":"8636db08b5425e5c464dc8f5aae692b68bdab14b9b01bcd67b6031e570796fa6"} Mar 21 05:29:50 crc kubenswrapper[4775]: I0321 05:29:50.186903 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" event={"ID":"bccbefa9-966d-44b9-bd8f-bb566649b315","Type":"ContainerStarted","Data":"f29e85380399c91a14fac8a091af8cf23a0384a224751b0b587f895a3d44aec1"} Mar 21 05:29:50 crc kubenswrapper[4775]: I0321 05:29:50.213348 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" podStartSLOduration=1.999674274 podStartE2EDuration="3.213326297s" podCreationTimestamp="2026-03-21 05:29:47 +0000 UTC" firstStartedPulling="2026-03-21 05:29:48.179520399 +0000 UTC m=+2541.155984023" lastFinishedPulling="2026-03-21 05:29:49.393172412 +0000 UTC m=+2542.369636046" observedRunningTime="2026-03-21 05:29:50.21237066 +0000 UTC m=+2543.188834284" watchObservedRunningTime="2026-03-21 05:29:50.213326297 +0000 UTC m=+2543.189789911" Mar 21 05:29:55 crc kubenswrapper[4775]: I0321 05:29:55.662212 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:29:55 crc kubenswrapper[4775]: E0321 05:29:55.663488 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.163780 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567850-gfbv5"] Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.166372 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567850-gfbv5" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.168964 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.172181 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.172331 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.176420 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567850-mp2j9"] Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.178995 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567850-mp2j9" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.181288 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.181695 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.184430 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bf67\" (UniqueName: \"kubernetes.io/projected/2616ee6e-f4b6-4258-b738-95ad6aab1644-kube-api-access-7bf67\") pod \"auto-csr-approver-29567850-gfbv5\" (UID: \"2616ee6e-f4b6-4258-b738-95ad6aab1644\") " pod="openshift-infra/auto-csr-approver-29567850-gfbv5" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.187911 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567850-mp2j9"] Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.200060 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567850-gfbv5"] Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.286283 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3215b5b-802c-4a03-b546-8e3ebe5f60b2-secret-volume\") pod \"collect-profiles-29567850-mp2j9\" (UID: \"a3215b5b-802c-4a03-b546-8e3ebe5f60b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567850-mp2j9" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.286367 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bf67\" (UniqueName: \"kubernetes.io/projected/2616ee6e-f4b6-4258-b738-95ad6aab1644-kube-api-access-7bf67\") pod \"auto-csr-approver-29567850-gfbv5\" (UID: \"2616ee6e-f4b6-4258-b738-95ad6aab1644\") " pod="openshift-infra/auto-csr-approver-29567850-gfbv5" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.286395 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3215b5b-802c-4a03-b546-8e3ebe5f60b2-config-volume\") pod \"collect-profiles-29567850-mp2j9\" (UID: \"a3215b5b-802c-4a03-b546-8e3ebe5f60b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567850-mp2j9" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.286473 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrsmx\" (UniqueName: \"kubernetes.io/projected/a3215b5b-802c-4a03-b546-8e3ebe5f60b2-kube-api-access-qrsmx\") pod \"collect-profiles-29567850-mp2j9\" (UID: \"a3215b5b-802c-4a03-b546-8e3ebe5f60b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567850-mp2j9" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.307238 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bf67\" (UniqueName: \"kubernetes.io/projected/2616ee6e-f4b6-4258-b738-95ad6aab1644-kube-api-access-7bf67\") pod \"auto-csr-approver-29567850-gfbv5\" (UID: \"2616ee6e-f4b6-4258-b738-95ad6aab1644\") " pod="openshift-infra/auto-csr-approver-29567850-gfbv5" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.388815 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3215b5b-802c-4a03-b546-8e3ebe5f60b2-secret-volume\") pod \"collect-profiles-29567850-mp2j9\" (UID: \"a3215b5b-802c-4a03-b546-8e3ebe5f60b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567850-mp2j9" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.389273 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3215b5b-802c-4a03-b546-8e3ebe5f60b2-config-volume\") pod \"collect-profiles-29567850-mp2j9\" (UID: \"a3215b5b-802c-4a03-b546-8e3ebe5f60b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567850-mp2j9" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.389327 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrsmx\" (UniqueName: \"kubernetes.io/projected/a3215b5b-802c-4a03-b546-8e3ebe5f60b2-kube-api-access-qrsmx\") pod \"collect-profiles-29567850-mp2j9\" (UID: \"a3215b5b-802c-4a03-b546-8e3ebe5f60b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567850-mp2j9" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.391163 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3215b5b-802c-4a03-b546-8e3ebe5f60b2-config-volume\") pod \"collect-profiles-29567850-mp2j9\" (UID: \"a3215b5b-802c-4a03-b546-8e3ebe5f60b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567850-mp2j9" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.395778 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3215b5b-802c-4a03-b546-8e3ebe5f60b2-secret-volume\") pod \"collect-profiles-29567850-mp2j9\" (UID: \"a3215b5b-802c-4a03-b546-8e3ebe5f60b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567850-mp2j9" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.412668 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrsmx\" (UniqueName: \"kubernetes.io/projected/a3215b5b-802c-4a03-b546-8e3ebe5f60b2-kube-api-access-qrsmx\") pod \"collect-profiles-29567850-mp2j9\" (UID: \"a3215b5b-802c-4a03-b546-8e3ebe5f60b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567850-mp2j9" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.493977 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567850-gfbv5" Mar 21 05:30:00 crc kubenswrapper[4775]: I0321 05:30:00.505736 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567850-mp2j9" Mar 21 05:30:01 crc kubenswrapper[4775]: I0321 05:30:01.019289 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567850-gfbv5"] Mar 21 05:30:01 crc kubenswrapper[4775]: I0321 05:30:01.119640 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567850-mp2j9"] Mar 21 05:30:01 crc kubenswrapper[4775]: W0321 05:30:01.122800 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3215b5b_802c_4a03_b546_8e3ebe5f60b2.slice/crio-d90243f8615cf1c263b1b17b7bb8bda9e62e280a115b1f6d61465f1d0981e792 WatchSource:0}: Error finding container d90243f8615cf1c263b1b17b7bb8bda9e62e280a115b1f6d61465f1d0981e792: Status 404 returned error can't find the container with id d90243f8615cf1c263b1b17b7bb8bda9e62e280a115b1f6d61465f1d0981e792 Mar 21 05:30:01 crc kubenswrapper[4775]: I0321 05:30:01.350916 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567850-gfbv5" event={"ID":"2616ee6e-f4b6-4258-b738-95ad6aab1644","Type":"ContainerStarted","Data":"5d4a88f575940f0e3ab1a15b409729e8419ee9828f39ed8e30ed6a9f8a6eb5f0"} Mar 21 05:30:01 crc kubenswrapper[4775]: I0321 05:30:01.353053 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567850-mp2j9" event={"ID":"a3215b5b-802c-4a03-b546-8e3ebe5f60b2","Type":"ContainerStarted","Data":"d90243f8615cf1c263b1b17b7bb8bda9e62e280a115b1f6d61465f1d0981e792"} Mar 21 05:30:02 crc kubenswrapper[4775]: I0321 05:30:02.366973 4775 generic.go:334] "Generic (PLEG): container finished" podID="a3215b5b-802c-4a03-b546-8e3ebe5f60b2" containerID="e0f13d79ecf702c5e141a30fa8c4b5464aeca1ab255956281f9d561257381418" exitCode=0 Mar 21 05:30:02 crc kubenswrapper[4775]: I0321 05:30:02.367435 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567850-mp2j9" event={"ID":"a3215b5b-802c-4a03-b546-8e3ebe5f60b2","Type":"ContainerDied","Data":"e0f13d79ecf702c5e141a30fa8c4b5464aeca1ab255956281f9d561257381418"} Mar 21 05:30:03 crc kubenswrapper[4775]: I0321 05:30:03.765808 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567850-mp2j9" Mar 21 05:30:03 crc kubenswrapper[4775]: I0321 05:30:03.872009 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrsmx\" (UniqueName: \"kubernetes.io/projected/a3215b5b-802c-4a03-b546-8e3ebe5f60b2-kube-api-access-qrsmx\") pod \"a3215b5b-802c-4a03-b546-8e3ebe5f60b2\" (UID: \"a3215b5b-802c-4a03-b546-8e3ebe5f60b2\") " Mar 21 05:30:03 crc kubenswrapper[4775]: I0321 05:30:03.872374 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3215b5b-802c-4a03-b546-8e3ebe5f60b2-config-volume\") pod \"a3215b5b-802c-4a03-b546-8e3ebe5f60b2\" (UID: \"a3215b5b-802c-4a03-b546-8e3ebe5f60b2\") " Mar 21 05:30:03 crc kubenswrapper[4775]: I0321 05:30:03.872582 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3215b5b-802c-4a03-b546-8e3ebe5f60b2-secret-volume\") pod \"a3215b5b-802c-4a03-b546-8e3ebe5f60b2\" (UID: \"a3215b5b-802c-4a03-b546-8e3ebe5f60b2\") " Mar 21 05:30:03 crc kubenswrapper[4775]: I0321 05:30:03.873330 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3215b5b-802c-4a03-b546-8e3ebe5f60b2-config-volume" (OuterVolumeSpecName: "config-volume") pod "a3215b5b-802c-4a03-b546-8e3ebe5f60b2" (UID: "a3215b5b-802c-4a03-b546-8e3ebe5f60b2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:30:03 crc kubenswrapper[4775]: I0321 05:30:03.881471 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3215b5b-802c-4a03-b546-8e3ebe5f60b2-kube-api-access-qrsmx" (OuterVolumeSpecName: "kube-api-access-qrsmx") pod "a3215b5b-802c-4a03-b546-8e3ebe5f60b2" (UID: "a3215b5b-802c-4a03-b546-8e3ebe5f60b2"). InnerVolumeSpecName "kube-api-access-qrsmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:30:03 crc kubenswrapper[4775]: I0321 05:30:03.896463 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3215b5b-802c-4a03-b546-8e3ebe5f60b2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a3215b5b-802c-4a03-b546-8e3ebe5f60b2" (UID: "a3215b5b-802c-4a03-b546-8e3ebe5f60b2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:30:03 crc kubenswrapper[4775]: I0321 05:30:03.975690 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3215b5b-802c-4a03-b546-8e3ebe5f60b2-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 21 05:30:03 crc kubenswrapper[4775]: I0321 05:30:03.976109 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrsmx\" (UniqueName: \"kubernetes.io/projected/a3215b5b-802c-4a03-b546-8e3ebe5f60b2-kube-api-access-qrsmx\") on node \"crc\" DevicePath \"\"" Mar 21 05:30:03 crc kubenswrapper[4775]: I0321 05:30:03.976194 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3215b5b-802c-4a03-b546-8e3ebe5f60b2-config-volume\") on node \"crc\" DevicePath \"\"" Mar 21 05:30:04 crc kubenswrapper[4775]: I0321 05:30:04.401311 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567850-mp2j9" event={"ID":"a3215b5b-802c-4a03-b546-8e3ebe5f60b2","Type":"ContainerDied","Data":"d90243f8615cf1c263b1b17b7bb8bda9e62e280a115b1f6d61465f1d0981e792"} Mar 21 05:30:04 crc kubenswrapper[4775]: I0321 05:30:04.401404 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d90243f8615cf1c263b1b17b7bb8bda9e62e280a115b1f6d61465f1d0981e792" Mar 21 05:30:04 crc kubenswrapper[4775]: I0321 05:30:04.401603 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567850-mp2j9" Mar 21 05:30:04 crc kubenswrapper[4775]: I0321 05:30:04.404370 4775 generic.go:334] "Generic (PLEG): container finished" podID="2616ee6e-f4b6-4258-b738-95ad6aab1644" containerID="2262a4a2de084f40220b6b85086a44442e4d874dfee1c84bf3dcc589fdc32198" exitCode=0 Mar 21 05:30:04 crc kubenswrapper[4775]: I0321 05:30:04.404410 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567850-gfbv5" event={"ID":"2616ee6e-f4b6-4258-b738-95ad6aab1644","Type":"ContainerDied","Data":"2262a4a2de084f40220b6b85086a44442e4d874dfee1c84bf3dcc589fdc32198"} Mar 21 05:30:04 crc kubenswrapper[4775]: I0321 05:30:04.851429 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct"] Mar 21 05:30:04 crc kubenswrapper[4775]: I0321 05:30:04.859468 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567805-97bct"] Mar 21 05:30:05 crc kubenswrapper[4775]: I0321 05:30:05.675297 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5aa6958-e573-4efb-a031-218c62b0bec9" path="/var/lib/kubelet/pods/a5aa6958-e573-4efb-a031-218c62b0bec9/volumes" Mar 21 05:30:05 crc kubenswrapper[4775]: I0321 05:30:05.798926 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567850-gfbv5" Mar 21 05:30:05 crc kubenswrapper[4775]: I0321 05:30:05.954262 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bf67\" (UniqueName: \"kubernetes.io/projected/2616ee6e-f4b6-4258-b738-95ad6aab1644-kube-api-access-7bf67\") pod \"2616ee6e-f4b6-4258-b738-95ad6aab1644\" (UID: \"2616ee6e-f4b6-4258-b738-95ad6aab1644\") " Mar 21 05:30:05 crc kubenswrapper[4775]: I0321 05:30:05.967505 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2616ee6e-f4b6-4258-b738-95ad6aab1644-kube-api-access-7bf67" (OuterVolumeSpecName: "kube-api-access-7bf67") pod "2616ee6e-f4b6-4258-b738-95ad6aab1644" (UID: "2616ee6e-f4b6-4258-b738-95ad6aab1644"). InnerVolumeSpecName "kube-api-access-7bf67". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:30:06 crc kubenswrapper[4775]: I0321 05:30:06.056744 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bf67\" (UniqueName: \"kubernetes.io/projected/2616ee6e-f4b6-4258-b738-95ad6aab1644-kube-api-access-7bf67\") on node \"crc\" DevicePath \"\"" Mar 21 05:30:06 crc kubenswrapper[4775]: I0321 05:30:06.430977 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567850-gfbv5" event={"ID":"2616ee6e-f4b6-4258-b738-95ad6aab1644","Type":"ContainerDied","Data":"5d4a88f575940f0e3ab1a15b409729e8419ee9828f39ed8e30ed6a9f8a6eb5f0"} Mar 21 05:30:06 crc kubenswrapper[4775]: I0321 05:30:06.431491 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d4a88f575940f0e3ab1a15b409729e8419ee9828f39ed8e30ed6a9f8a6eb5f0" Mar 21 05:30:06 crc kubenswrapper[4775]: I0321 05:30:06.431057 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567850-gfbv5" Mar 21 05:30:06 crc kubenswrapper[4775]: I0321 05:30:06.871636 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567844-25mvs"] Mar 21 05:30:06 crc kubenswrapper[4775]: I0321 05:30:06.885170 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567844-25mvs"] Mar 21 05:30:07 crc kubenswrapper[4775]: I0321 05:30:07.673704 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6" path="/var/lib/kubelet/pods/c589176d-45ef-4e7a-b8f8-0ec31e4fe6a6/volumes" Mar 21 05:30:08 crc kubenswrapper[4775]: I0321 05:30:08.662771 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:30:08 crc kubenswrapper[4775]: E0321 05:30:08.663107 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:30:21 crc kubenswrapper[4775]: I0321 05:30:21.661808 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:30:21 crc kubenswrapper[4775]: E0321 05:30:21.662833 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:30:35 crc kubenswrapper[4775]: I0321 05:30:35.662405 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:30:36 crc kubenswrapper[4775]: I0321 05:30:36.728031 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerStarted","Data":"ea40506310b7606d65a9c28447427b7d1714319295a97c7d0f6f7a3b4b8251ee"} Mar 21 05:30:47 crc kubenswrapper[4775]: I0321 05:30:47.611990 4775 scope.go:117] "RemoveContainer" containerID="5936961592e0ba9da2bfbec1bc18d0df5f38a71ab49e8c90edaa7dea4ccb02ee" Mar 21 05:30:47 crc kubenswrapper[4775]: I0321 05:30:47.669875 4775 scope.go:117] "RemoveContainer" containerID="56e35510c790be074d09f5ee8f124d413aaf4e647a0fe13993e83a909111b979" Mar 21 05:30:47 crc kubenswrapper[4775]: I0321 05:30:47.702042 4775 scope.go:117] "RemoveContainer" containerID="1b93f4361b64d716502128254e9ea25b78ba9ed772325baf5c04d50a5a77ab40" Mar 21 05:31:47 crc kubenswrapper[4775]: I0321 05:31:47.808901 4775 scope.go:117] "RemoveContainer" containerID="60fb5be3131696012c97db115a74ff472115d80eb7fd9153e9102904c69c91b3" Mar 21 05:31:47 crc kubenswrapper[4775]: I0321 05:31:47.843259 4775 scope.go:117] "RemoveContainer" containerID="a2a8c8dc1d88f43d6ededf2e8793a512b5087dcbf7e313f245d34485e136258b" Mar 21 05:32:00 crc kubenswrapper[4775]: I0321 05:32:00.154137 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567852-hd7nd"] Mar 21 05:32:00 crc kubenswrapper[4775]: E0321 05:32:00.155615 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2616ee6e-f4b6-4258-b738-95ad6aab1644" containerName="oc" Mar 21 05:32:00 crc kubenswrapper[4775]: I0321 05:32:00.155635 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2616ee6e-f4b6-4258-b738-95ad6aab1644" containerName="oc" Mar 21 05:32:00 crc kubenswrapper[4775]: E0321 05:32:00.155684 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3215b5b-802c-4a03-b546-8e3ebe5f60b2" containerName="collect-profiles" Mar 21 05:32:00 crc kubenswrapper[4775]: I0321 05:32:00.155692 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3215b5b-802c-4a03-b546-8e3ebe5f60b2" containerName="collect-profiles" Mar 21 05:32:00 crc kubenswrapper[4775]: I0321 05:32:00.156009 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2616ee6e-f4b6-4258-b738-95ad6aab1644" containerName="oc" Mar 21 05:32:00 crc kubenswrapper[4775]: I0321 05:32:00.156030 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3215b5b-802c-4a03-b546-8e3ebe5f60b2" containerName="collect-profiles" Mar 21 05:32:00 crc kubenswrapper[4775]: I0321 05:32:00.157164 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567852-hd7nd" Mar 21 05:32:00 crc kubenswrapper[4775]: I0321 05:32:00.163662 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:32:00 crc kubenswrapper[4775]: I0321 05:32:00.163831 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:32:00 crc kubenswrapper[4775]: I0321 05:32:00.164233 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:32:00 crc kubenswrapper[4775]: I0321 05:32:00.166771 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567852-hd7nd"] Mar 21 05:32:00 crc kubenswrapper[4775]: I0321 05:32:00.216827 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xk4k\" (UniqueName: \"kubernetes.io/projected/15017759-960b-4d6e-8ce9-deef2cb94155-kube-api-access-4xk4k\") pod \"auto-csr-approver-29567852-hd7nd\" (UID: \"15017759-960b-4d6e-8ce9-deef2cb94155\") " pod="openshift-infra/auto-csr-approver-29567852-hd7nd" Mar 21 05:32:00 crc kubenswrapper[4775]: I0321 05:32:00.318984 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xk4k\" (UniqueName: \"kubernetes.io/projected/15017759-960b-4d6e-8ce9-deef2cb94155-kube-api-access-4xk4k\") pod \"auto-csr-approver-29567852-hd7nd\" (UID: \"15017759-960b-4d6e-8ce9-deef2cb94155\") " pod="openshift-infra/auto-csr-approver-29567852-hd7nd" Mar 21 05:32:00 crc kubenswrapper[4775]: I0321 05:32:00.346908 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xk4k\" (UniqueName: \"kubernetes.io/projected/15017759-960b-4d6e-8ce9-deef2cb94155-kube-api-access-4xk4k\") pod \"auto-csr-approver-29567852-hd7nd\" (UID: \"15017759-960b-4d6e-8ce9-deef2cb94155\") " pod="openshift-infra/auto-csr-approver-29567852-hd7nd" Mar 21 05:32:00 crc kubenswrapper[4775]: I0321 05:32:00.483198 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567852-hd7nd" Mar 21 05:32:01 crc kubenswrapper[4775]: I0321 05:32:01.007670 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567852-hd7nd"] Mar 21 05:32:01 crc kubenswrapper[4775]: I0321 05:32:01.014546 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 05:32:01 crc kubenswrapper[4775]: I0321 05:32:01.636163 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567852-hd7nd" event={"ID":"15017759-960b-4d6e-8ce9-deef2cb94155","Type":"ContainerStarted","Data":"6a3a151d48881d011e498b74117ddba9087c8a3bccd9ea12c290dab570221f0c"} Mar 21 05:32:02 crc kubenswrapper[4775]: I0321 05:32:02.651338 4775 generic.go:334] "Generic (PLEG): container finished" podID="15017759-960b-4d6e-8ce9-deef2cb94155" containerID="25828aaca134034962b25d7ea45e1870919d75c0c4923758d122b6c7bce2397c" exitCode=0 Mar 21 05:32:02 crc kubenswrapper[4775]: I0321 05:32:02.651466 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567852-hd7nd" event={"ID":"15017759-960b-4d6e-8ce9-deef2cb94155","Type":"ContainerDied","Data":"25828aaca134034962b25d7ea45e1870919d75c0c4923758d122b6c7bce2397c"} Mar 21 05:32:02 crc kubenswrapper[4775]: I0321 05:32:02.654622 4775 generic.go:334] "Generic (PLEG): container finished" podID="bccbefa9-966d-44b9-bd8f-bb566649b315" containerID="f29e85380399c91a14fac8a091af8cf23a0384a224751b0b587f895a3d44aec1" exitCode=0 Mar 21 05:32:02 crc kubenswrapper[4775]: I0321 05:32:02.654680 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" event={"ID":"bccbefa9-966d-44b9-bd8f-bb566649b315","Type":"ContainerDied","Data":"f29e85380399c91a14fac8a091af8cf23a0384a224751b0b587f895a3d44aec1"} Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.164145 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567852-hd7nd" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.170746 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.325590 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pskfk\" (UniqueName: \"kubernetes.io/projected/bccbefa9-966d-44b9-bd8f-bb566649b315-kube-api-access-pskfk\") pod \"bccbefa9-966d-44b9-bd8f-bb566649b315\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.325799 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ceilometer-compute-config-data-2\") pod \"bccbefa9-966d-44b9-bd8f-bb566649b315\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.325844 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xk4k\" (UniqueName: \"kubernetes.io/projected/15017759-960b-4d6e-8ce9-deef2cb94155-kube-api-access-4xk4k\") pod \"15017759-960b-4d6e-8ce9-deef2cb94155\" (UID: \"15017759-960b-4d6e-8ce9-deef2cb94155\") " Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.325927 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ssh-key-openstack-edpm-ipam\") pod \"bccbefa9-966d-44b9-bd8f-bb566649b315\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.326040 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-telemetry-combined-ca-bundle\") pod \"bccbefa9-966d-44b9-bd8f-bb566649b315\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.326099 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ceilometer-compute-config-data-0\") pod \"bccbefa9-966d-44b9-bd8f-bb566649b315\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.326172 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-inventory\") pod \"bccbefa9-966d-44b9-bd8f-bb566649b315\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.326243 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ceilometer-compute-config-data-1\") pod \"bccbefa9-966d-44b9-bd8f-bb566649b315\" (UID: \"bccbefa9-966d-44b9-bd8f-bb566649b315\") " Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.335654 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bccbefa9-966d-44b9-bd8f-bb566649b315-kube-api-access-pskfk" (OuterVolumeSpecName: "kube-api-access-pskfk") pod "bccbefa9-966d-44b9-bd8f-bb566649b315" (UID: "bccbefa9-966d-44b9-bd8f-bb566649b315"). InnerVolumeSpecName "kube-api-access-pskfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.336544 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "bccbefa9-966d-44b9-bd8f-bb566649b315" (UID: "bccbefa9-966d-44b9-bd8f-bb566649b315"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.336791 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15017759-960b-4d6e-8ce9-deef2cb94155-kube-api-access-4xk4k" (OuterVolumeSpecName: "kube-api-access-4xk4k") pod "15017759-960b-4d6e-8ce9-deef2cb94155" (UID: "15017759-960b-4d6e-8ce9-deef2cb94155"). InnerVolumeSpecName "kube-api-access-4xk4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.361441 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "bccbefa9-966d-44b9-bd8f-bb566649b315" (UID: "bccbefa9-966d-44b9-bd8f-bb566649b315"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.369426 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "bccbefa9-966d-44b9-bd8f-bb566649b315" (UID: "bccbefa9-966d-44b9-bd8f-bb566649b315"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.369535 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-inventory" (OuterVolumeSpecName: "inventory") pod "bccbefa9-966d-44b9-bd8f-bb566649b315" (UID: "bccbefa9-966d-44b9-bd8f-bb566649b315"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.372963 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "bccbefa9-966d-44b9-bd8f-bb566649b315" (UID: "bccbefa9-966d-44b9-bd8f-bb566649b315"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.374435 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "bccbefa9-966d-44b9-bd8f-bb566649b315" (UID: "bccbefa9-966d-44b9-bd8f-bb566649b315"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.430222 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pskfk\" (UniqueName: \"kubernetes.io/projected/bccbefa9-966d-44b9-bd8f-bb566649b315-kube-api-access-pskfk\") on node \"crc\" DevicePath \"\"" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.430352 4775 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.430411 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xk4k\" (UniqueName: \"kubernetes.io/projected/15017759-960b-4d6e-8ce9-deef2cb94155-kube-api-access-4xk4k\") on node \"crc\" DevicePath \"\"" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.430468 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.430520 4775 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.430587 4775 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.430645 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.430706 4775 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/bccbefa9-966d-44b9-bd8f-bb566649b315-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.675657 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" event={"ID":"bccbefa9-966d-44b9-bd8f-bb566649b315","Type":"ContainerDied","Data":"8636db08b5425e5c464dc8f5aae692b68bdab14b9b01bcd67b6031e570796fa6"} Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.675747 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8636db08b5425e5c464dc8f5aae692b68bdab14b9b01bcd67b6031e570796fa6" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.675752 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.677507 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567852-hd7nd" event={"ID":"15017759-960b-4d6e-8ce9-deef2cb94155","Type":"ContainerDied","Data":"6a3a151d48881d011e498b74117ddba9087c8a3bccd9ea12c290dab570221f0c"} Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.677551 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a3a151d48881d011e498b74117ddba9087c8a3bccd9ea12c290dab570221f0c" Mar 21 05:32:04 crc kubenswrapper[4775]: I0321 05:32:04.677616 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567852-hd7nd" Mar 21 05:32:05 crc kubenswrapper[4775]: I0321 05:32:05.259558 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567846-6lslz"] Mar 21 05:32:05 crc kubenswrapper[4775]: I0321 05:32:05.274251 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567846-6lslz"] Mar 21 05:32:05 crc kubenswrapper[4775]: I0321 05:32:05.673113 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43f89ba5-2605-4bdb-bc77-020c6bf9db75" path="/var/lib/kubelet/pods/43f89ba5-2605-4bdb-bc77-020c6bf9db75/volumes" Mar 21 05:32:33 crc kubenswrapper[4775]: E0321 05:32:33.182775 4775 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.200:36540->38.102.83.200:42597: read tcp 38.102.83.200:36540->38.102.83.200:42597: read: connection reset by peer Mar 21 05:32:47 crc kubenswrapper[4775]: I0321 05:32:47.911996 4775 scope.go:117] "RemoveContainer" containerID="a8931f9c5d17f0e3f84a26748f81ac4ac575c43d3a88da2c5e6613b2fee09eee" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.217909 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Mar 21 05:32:59 crc kubenswrapper[4775]: E0321 05:32:59.219040 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bccbefa9-966d-44b9-bd8f-bb566649b315" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.219057 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="bccbefa9-966d-44b9-bd8f-bb566649b315" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 21 05:32:59 crc kubenswrapper[4775]: E0321 05:32:59.219082 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15017759-960b-4d6e-8ce9-deef2cb94155" containerName="oc" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.219089 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="15017759-960b-4d6e-8ce9-deef2cb94155" containerName="oc" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.219287 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="15017759-960b-4d6e-8ce9-deef2cb94155" containerName="oc" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.219303 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="bccbefa9-966d-44b9-bd8f-bb566649b315" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.220002 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.223302 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.223698 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.225093 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.226406 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-bnv6q" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.246025 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.279600 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1c832898-838d-423d-8ad8-512c5ee5706c-config-data\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.279728 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1c832898-838d-423d-8ad8-512c5ee5706c-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.279869 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1c832898-838d-423d-8ad8-512c5ee5706c-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.381546 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1c832898-838d-423d-8ad8-512c5ee5706c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.381623 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1c832898-838d-423d-8ad8-512c5ee5706c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.381668 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1c832898-838d-423d-8ad8-512c5ee5706c-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.381728 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1c832898-838d-423d-8ad8-512c5ee5706c-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.381784 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1c832898-838d-423d-8ad8-512c5ee5706c-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.381824 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1c832898-838d-423d-8ad8-512c5ee5706c-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.381872 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdw2v\" (UniqueName: \"kubernetes.io/projected/1c832898-838d-423d-8ad8-512c5ee5706c-kube-api-access-gdw2v\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.381912 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1c832898-838d-423d-8ad8-512c5ee5706c-config-data\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.381946 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.384146 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1c832898-838d-423d-8ad8-512c5ee5706c-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.385282 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1c832898-838d-423d-8ad8-512c5ee5706c-config-data\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.400324 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1c832898-838d-423d-8ad8-512c5ee5706c-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.484092 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1c832898-838d-423d-8ad8-512c5ee5706c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.484186 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1c832898-838d-423d-8ad8-512c5ee5706c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.484286 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1c832898-838d-423d-8ad8-512c5ee5706c-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.484362 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1c832898-838d-423d-8ad8-512c5ee5706c-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.484463 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdw2v\" (UniqueName: \"kubernetes.io/projected/1c832898-838d-423d-8ad8-512c5ee5706c-kube-api-access-gdw2v\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.484550 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.484594 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1c832898-838d-423d-8ad8-512c5ee5706c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.484661 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1c832898-838d-423d-8ad8-512c5ee5706c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.485066 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.489430 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1c832898-838d-423d-8ad8-512c5ee5706c-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.490462 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1c832898-838d-423d-8ad8-512c5ee5706c-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.505455 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdw2v\" (UniqueName: \"kubernetes.io/projected/1c832898-838d-423d-8ad8-512c5ee5706c-kube-api-access-gdw2v\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.518317 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " pod="openstack/tempest-tests-tempest" Mar 21 05:32:59 crc kubenswrapper[4775]: I0321 05:32:59.552965 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 21 05:33:00 crc kubenswrapper[4775]: I0321 05:33:00.086047 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Mar 21 05:33:00 crc kubenswrapper[4775]: I0321 05:33:00.305058 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1c832898-838d-423d-8ad8-512c5ee5706c","Type":"ContainerStarted","Data":"695de001f7207ca22144c3ce97e3ea6306a89b43020987a1d4856f3f9e289416"} Mar 21 05:33:02 crc kubenswrapper[4775]: I0321 05:33:02.482249 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:33:02 crc kubenswrapper[4775]: I0321 05:33:02.483169 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:33:12 crc kubenswrapper[4775]: E0321 05:33:12.008706 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Mar 21 05:33:31 crc kubenswrapper[4775]: E0321 05:33:31.349343 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Mar 21 05:33:31 crc kubenswrapper[4775]: E0321 05:33:31.350316 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gdw2v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(1c832898-838d-423d-8ad8-512c5ee5706c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 05:33:31 crc kubenswrapper[4775]: E0321 05:33:31.351792 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="1c832898-838d-423d-8ad8-512c5ee5706c" Mar 21 05:33:31 crc kubenswrapper[4775]: E0321 05:33:31.776680 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="1c832898-838d-423d-8ad8-512c5ee5706c" Mar 21 05:33:32 crc kubenswrapper[4775]: I0321 05:33:32.482084 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:33:32 crc kubenswrapper[4775]: I0321 05:33:32.482555 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:33:45 crc kubenswrapper[4775]: I0321 05:33:45.426854 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Mar 21 05:33:46 crc kubenswrapper[4775]: I0321 05:33:46.945670 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1c832898-838d-423d-8ad8-512c5ee5706c","Type":"ContainerStarted","Data":"e5d5a6749ca728ee30628ed8c23d0df8c8193d878227278bf224646abd9c8806"} Mar 21 05:33:46 crc kubenswrapper[4775]: I0321 05:33:46.972863 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.6497522570000003 podStartE2EDuration="48.972841682s" podCreationTimestamp="2026-03-21 05:32:58 +0000 UTC" firstStartedPulling="2026-03-21 05:33:00.099423383 +0000 UTC m=+2733.075887047" lastFinishedPulling="2026-03-21 05:33:45.422512798 +0000 UTC m=+2778.398976472" observedRunningTime="2026-03-21 05:33:46.969324291 +0000 UTC m=+2779.945787925" watchObservedRunningTime="2026-03-21 05:33:46.972841682 +0000 UTC m=+2779.949305306" Mar 21 05:33:50 crc kubenswrapper[4775]: I0321 05:33:50.367616 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4chpg"] Mar 21 05:33:50 crc kubenswrapper[4775]: I0321 05:33:50.372667 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4chpg" Mar 21 05:33:50 crc kubenswrapper[4775]: I0321 05:33:50.388433 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4chpg"] Mar 21 05:33:50 crc kubenswrapper[4775]: I0321 05:33:50.485896 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/501afe19-8db4-4d25-8e06-56d1449b3643-utilities\") pod \"certified-operators-4chpg\" (UID: \"501afe19-8db4-4d25-8e06-56d1449b3643\") " pod="openshift-marketplace/certified-operators-4chpg" Mar 21 05:33:50 crc kubenswrapper[4775]: I0321 05:33:50.486092 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/501afe19-8db4-4d25-8e06-56d1449b3643-catalog-content\") pod \"certified-operators-4chpg\" (UID: \"501afe19-8db4-4d25-8e06-56d1449b3643\") " pod="openshift-marketplace/certified-operators-4chpg" Mar 21 05:33:50 crc kubenswrapper[4775]: I0321 05:33:50.486204 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q46w\" (UniqueName: \"kubernetes.io/projected/501afe19-8db4-4d25-8e06-56d1449b3643-kube-api-access-8q46w\") pod \"certified-operators-4chpg\" (UID: \"501afe19-8db4-4d25-8e06-56d1449b3643\") " pod="openshift-marketplace/certified-operators-4chpg" Mar 21 05:33:50 crc kubenswrapper[4775]: I0321 05:33:50.588135 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q46w\" (UniqueName: \"kubernetes.io/projected/501afe19-8db4-4d25-8e06-56d1449b3643-kube-api-access-8q46w\") pod \"certified-operators-4chpg\" (UID: \"501afe19-8db4-4d25-8e06-56d1449b3643\") " pod="openshift-marketplace/certified-operators-4chpg" Mar 21 05:33:50 crc kubenswrapper[4775]: I0321 05:33:50.588240 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/501afe19-8db4-4d25-8e06-56d1449b3643-utilities\") pod \"certified-operators-4chpg\" (UID: \"501afe19-8db4-4d25-8e06-56d1449b3643\") " pod="openshift-marketplace/certified-operators-4chpg" Mar 21 05:33:50 crc kubenswrapper[4775]: I0321 05:33:50.588361 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/501afe19-8db4-4d25-8e06-56d1449b3643-catalog-content\") pod \"certified-operators-4chpg\" (UID: \"501afe19-8db4-4d25-8e06-56d1449b3643\") " pod="openshift-marketplace/certified-operators-4chpg" Mar 21 05:33:50 crc kubenswrapper[4775]: I0321 05:33:50.588977 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/501afe19-8db4-4d25-8e06-56d1449b3643-catalog-content\") pod \"certified-operators-4chpg\" (UID: \"501afe19-8db4-4d25-8e06-56d1449b3643\") " pod="openshift-marketplace/certified-operators-4chpg" Mar 21 05:33:50 crc kubenswrapper[4775]: I0321 05:33:50.589613 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/501afe19-8db4-4d25-8e06-56d1449b3643-utilities\") pod \"certified-operators-4chpg\" (UID: \"501afe19-8db4-4d25-8e06-56d1449b3643\") " pod="openshift-marketplace/certified-operators-4chpg" Mar 21 05:33:50 crc kubenswrapper[4775]: I0321 05:33:50.612179 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q46w\" (UniqueName: \"kubernetes.io/projected/501afe19-8db4-4d25-8e06-56d1449b3643-kube-api-access-8q46w\") pod \"certified-operators-4chpg\" (UID: \"501afe19-8db4-4d25-8e06-56d1449b3643\") " pod="openshift-marketplace/certified-operators-4chpg" Mar 21 05:33:50 crc kubenswrapper[4775]: I0321 05:33:50.701291 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4chpg" Mar 21 05:33:51 crc kubenswrapper[4775]: I0321 05:33:51.314668 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4chpg"] Mar 21 05:33:52 crc kubenswrapper[4775]: I0321 05:33:51.999601 4775 generic.go:334] "Generic (PLEG): container finished" podID="501afe19-8db4-4d25-8e06-56d1449b3643" containerID="4237a001ec9e64bac729ab0c016c015a46734930328f1c25302192aa4945b0bd" exitCode=0 Mar 21 05:33:52 crc kubenswrapper[4775]: I0321 05:33:51.999666 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4chpg" event={"ID":"501afe19-8db4-4d25-8e06-56d1449b3643","Type":"ContainerDied","Data":"4237a001ec9e64bac729ab0c016c015a46734930328f1c25302192aa4945b0bd"} Mar 21 05:33:52 crc kubenswrapper[4775]: I0321 05:33:52.000004 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4chpg" event={"ID":"501afe19-8db4-4d25-8e06-56d1449b3643","Type":"ContainerStarted","Data":"07eb8f05c11ebbbb5bb476c41bb59b876a26256bd2f154d72bba449f69c622df"} Mar 21 05:33:54 crc kubenswrapper[4775]: I0321 05:33:54.026233 4775 generic.go:334] "Generic (PLEG): container finished" podID="501afe19-8db4-4d25-8e06-56d1449b3643" containerID="c1908a0d86835431c41047756017337154557cff7b1ff87a8e9a3ad7ba39c029" exitCode=0 Mar 21 05:33:54 crc kubenswrapper[4775]: I0321 05:33:54.026296 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4chpg" event={"ID":"501afe19-8db4-4d25-8e06-56d1449b3643","Type":"ContainerDied","Data":"c1908a0d86835431c41047756017337154557cff7b1ff87a8e9a3ad7ba39c029"} Mar 21 05:33:55 crc kubenswrapper[4775]: I0321 05:33:55.042095 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4chpg" event={"ID":"501afe19-8db4-4d25-8e06-56d1449b3643","Type":"ContainerStarted","Data":"711a82a5ea8d72323dd172e53eaa58caf06be2ca6f1b2822acee6887a437466b"} Mar 21 05:33:55 crc kubenswrapper[4775]: I0321 05:33:55.072922 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4chpg" podStartSLOduration=2.584060545 podStartE2EDuration="5.072897873s" podCreationTimestamp="2026-03-21 05:33:50 +0000 UTC" firstStartedPulling="2026-03-21 05:33:52.006080577 +0000 UTC m=+2784.982544201" lastFinishedPulling="2026-03-21 05:33:54.494917905 +0000 UTC m=+2787.471381529" observedRunningTime="2026-03-21 05:33:55.062922335 +0000 UTC m=+2788.039385969" watchObservedRunningTime="2026-03-21 05:33:55.072897873 +0000 UTC m=+2788.049361497" Mar 21 05:34:00 crc kubenswrapper[4775]: I0321 05:34:00.146167 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567854-gxf82"] Mar 21 05:34:00 crc kubenswrapper[4775]: I0321 05:34:00.148172 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567854-gxf82" Mar 21 05:34:00 crc kubenswrapper[4775]: I0321 05:34:00.150883 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:34:00 crc kubenswrapper[4775]: I0321 05:34:00.150884 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:34:00 crc kubenswrapper[4775]: I0321 05:34:00.152321 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:34:00 crc kubenswrapper[4775]: I0321 05:34:00.158461 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567854-gxf82"] Mar 21 05:34:00 crc kubenswrapper[4775]: I0321 05:34:00.317045 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7kzr\" (UniqueName: \"kubernetes.io/projected/afdb9c39-b2aa-496c-8cfb-917fd0b1cc15-kube-api-access-b7kzr\") pod \"auto-csr-approver-29567854-gxf82\" (UID: \"afdb9c39-b2aa-496c-8cfb-917fd0b1cc15\") " pod="openshift-infra/auto-csr-approver-29567854-gxf82" Mar 21 05:34:00 crc kubenswrapper[4775]: I0321 05:34:00.420108 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7kzr\" (UniqueName: \"kubernetes.io/projected/afdb9c39-b2aa-496c-8cfb-917fd0b1cc15-kube-api-access-b7kzr\") pod \"auto-csr-approver-29567854-gxf82\" (UID: \"afdb9c39-b2aa-496c-8cfb-917fd0b1cc15\") " pod="openshift-infra/auto-csr-approver-29567854-gxf82" Mar 21 05:34:00 crc kubenswrapper[4775]: I0321 05:34:00.451192 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7kzr\" (UniqueName: \"kubernetes.io/projected/afdb9c39-b2aa-496c-8cfb-917fd0b1cc15-kube-api-access-b7kzr\") pod \"auto-csr-approver-29567854-gxf82\" (UID: \"afdb9c39-b2aa-496c-8cfb-917fd0b1cc15\") " pod="openshift-infra/auto-csr-approver-29567854-gxf82" Mar 21 05:34:00 crc kubenswrapper[4775]: I0321 05:34:00.474541 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567854-gxf82" Mar 21 05:34:00 crc kubenswrapper[4775]: I0321 05:34:00.702222 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4chpg" Mar 21 05:34:00 crc kubenswrapper[4775]: I0321 05:34:00.702596 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4chpg" Mar 21 05:34:00 crc kubenswrapper[4775]: I0321 05:34:00.777731 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4chpg" Mar 21 05:34:01 crc kubenswrapper[4775]: I0321 05:34:01.040104 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567854-gxf82"] Mar 21 05:34:01 crc kubenswrapper[4775]: I0321 05:34:01.107590 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567854-gxf82" event={"ID":"afdb9c39-b2aa-496c-8cfb-917fd0b1cc15","Type":"ContainerStarted","Data":"0d244af90f1f3cc12a466a4d5d229c239003d7b502e8be0d3f5edcd24ca33e5a"} Mar 21 05:34:01 crc kubenswrapper[4775]: I0321 05:34:01.161663 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4chpg" Mar 21 05:34:01 crc kubenswrapper[4775]: I0321 05:34:01.213955 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4chpg"] Mar 21 05:34:02 crc kubenswrapper[4775]: I0321 05:34:02.482840 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:34:02 crc kubenswrapper[4775]: I0321 05:34:02.483365 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:34:02 crc kubenswrapper[4775]: I0321 05:34:02.483414 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 05:34:02 crc kubenswrapper[4775]: I0321 05:34:02.485368 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ea40506310b7606d65a9c28447427b7d1714319295a97c7d0f6f7a3b4b8251ee"} pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 05:34:02 crc kubenswrapper[4775]: I0321 05:34:02.485420 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" containerID="cri-o://ea40506310b7606d65a9c28447427b7d1714319295a97c7d0f6f7a3b4b8251ee" gracePeriod=600 Mar 21 05:34:03 crc kubenswrapper[4775]: I0321 05:34:03.129897 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567854-gxf82" event={"ID":"afdb9c39-b2aa-496c-8cfb-917fd0b1cc15","Type":"ContainerStarted","Data":"20714a401828b9647e317e22423f1e191dc985cb13d154d0fa57a94896564896"} Mar 21 05:34:03 crc kubenswrapper[4775]: I0321 05:34:03.133093 4775 generic.go:334] "Generic (PLEG): container finished" podID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerID="ea40506310b7606d65a9c28447427b7d1714319295a97c7d0f6f7a3b4b8251ee" exitCode=0 Mar 21 05:34:03 crc kubenswrapper[4775]: I0321 05:34:03.133178 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerDied","Data":"ea40506310b7606d65a9c28447427b7d1714319295a97c7d0f6f7a3b4b8251ee"} Mar 21 05:34:03 crc kubenswrapper[4775]: I0321 05:34:03.133263 4775 scope.go:117] "RemoveContainer" containerID="21dc53d475648a02ba4409f119cfb163b347fd5f7ea3ee585b368e11bf55d804" Mar 21 05:34:03 crc kubenswrapper[4775]: I0321 05:34:03.133349 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4chpg" podUID="501afe19-8db4-4d25-8e06-56d1449b3643" containerName="registry-server" containerID="cri-o://711a82a5ea8d72323dd172e53eaa58caf06be2ca6f1b2822acee6887a437466b" gracePeriod=2 Mar 21 05:34:03 crc kubenswrapper[4775]: I0321 05:34:03.154019 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29567854-gxf82" podStartSLOduration=1.902136061 podStartE2EDuration="3.153998978s" podCreationTimestamp="2026-03-21 05:34:00 +0000 UTC" firstStartedPulling="2026-03-21 05:34:01.056372519 +0000 UTC m=+2794.032836143" lastFinishedPulling="2026-03-21 05:34:02.308235406 +0000 UTC m=+2795.284699060" observedRunningTime="2026-03-21 05:34:03.148666904 +0000 UTC m=+2796.125130548" watchObservedRunningTime="2026-03-21 05:34:03.153998978 +0000 UTC m=+2796.130462592" Mar 21 05:34:03 crc kubenswrapper[4775]: E0321 05:34:03.443709 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod501afe19_8db4_4d25_8e06_56d1449b3643.slice/crio-711a82a5ea8d72323dd172e53eaa58caf06be2ca6f1b2822acee6887a437466b.scope\": RecentStats: unable to find data in memory cache]" Mar 21 05:34:03 crc kubenswrapper[4775]: I0321 05:34:03.748084 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4chpg" Mar 21 05:34:03 crc kubenswrapper[4775]: I0321 05:34:03.922329 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q46w\" (UniqueName: \"kubernetes.io/projected/501afe19-8db4-4d25-8e06-56d1449b3643-kube-api-access-8q46w\") pod \"501afe19-8db4-4d25-8e06-56d1449b3643\" (UID: \"501afe19-8db4-4d25-8e06-56d1449b3643\") " Mar 21 05:34:03 crc kubenswrapper[4775]: I0321 05:34:03.923211 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/501afe19-8db4-4d25-8e06-56d1449b3643-catalog-content\") pod \"501afe19-8db4-4d25-8e06-56d1449b3643\" (UID: \"501afe19-8db4-4d25-8e06-56d1449b3643\") " Mar 21 05:34:03 crc kubenswrapper[4775]: I0321 05:34:03.923268 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/501afe19-8db4-4d25-8e06-56d1449b3643-utilities\") pod \"501afe19-8db4-4d25-8e06-56d1449b3643\" (UID: \"501afe19-8db4-4d25-8e06-56d1449b3643\") " Mar 21 05:34:03 crc kubenswrapper[4775]: I0321 05:34:03.924846 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/501afe19-8db4-4d25-8e06-56d1449b3643-utilities" (OuterVolumeSpecName: "utilities") pod "501afe19-8db4-4d25-8e06-56d1449b3643" (UID: "501afe19-8db4-4d25-8e06-56d1449b3643"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:34:03 crc kubenswrapper[4775]: I0321 05:34:03.947201 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/501afe19-8db4-4d25-8e06-56d1449b3643-kube-api-access-8q46w" (OuterVolumeSpecName: "kube-api-access-8q46w") pod "501afe19-8db4-4d25-8e06-56d1449b3643" (UID: "501afe19-8db4-4d25-8e06-56d1449b3643"). InnerVolumeSpecName "kube-api-access-8q46w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:34:03 crc kubenswrapper[4775]: I0321 05:34:03.988518 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/501afe19-8db4-4d25-8e06-56d1449b3643-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "501afe19-8db4-4d25-8e06-56d1449b3643" (UID: "501afe19-8db4-4d25-8e06-56d1449b3643"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.026527 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/501afe19-8db4-4d25-8e06-56d1449b3643-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.026575 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q46w\" (UniqueName: \"kubernetes.io/projected/501afe19-8db4-4d25-8e06-56d1449b3643-kube-api-access-8q46w\") on node \"crc\" DevicePath \"\"" Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.026588 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/501afe19-8db4-4d25-8e06-56d1449b3643-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.148162 4775 generic.go:334] "Generic (PLEG): container finished" podID="afdb9c39-b2aa-496c-8cfb-917fd0b1cc15" containerID="20714a401828b9647e317e22423f1e191dc985cb13d154d0fa57a94896564896" exitCode=0 Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.148263 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567854-gxf82" event={"ID":"afdb9c39-b2aa-496c-8cfb-917fd0b1cc15","Type":"ContainerDied","Data":"20714a401828b9647e317e22423f1e191dc985cb13d154d0fa57a94896564896"} Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.153160 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerStarted","Data":"f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9"} Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.158159 4775 generic.go:334] "Generic (PLEG): container finished" podID="501afe19-8db4-4d25-8e06-56d1449b3643" containerID="711a82a5ea8d72323dd172e53eaa58caf06be2ca6f1b2822acee6887a437466b" exitCode=0 Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.158216 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4chpg" event={"ID":"501afe19-8db4-4d25-8e06-56d1449b3643","Type":"ContainerDied","Data":"711a82a5ea8d72323dd172e53eaa58caf06be2ca6f1b2822acee6887a437466b"} Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.158254 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4chpg" event={"ID":"501afe19-8db4-4d25-8e06-56d1449b3643","Type":"ContainerDied","Data":"07eb8f05c11ebbbb5bb476c41bb59b876a26256bd2f154d72bba449f69c622df"} Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.158277 4775 scope.go:117] "RemoveContainer" containerID="711a82a5ea8d72323dd172e53eaa58caf06be2ca6f1b2822acee6887a437466b" Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.158277 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4chpg" Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.187549 4775 scope.go:117] "RemoveContainer" containerID="c1908a0d86835431c41047756017337154557cff7b1ff87a8e9a3ad7ba39c029" Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.223484 4775 scope.go:117] "RemoveContainer" containerID="4237a001ec9e64bac729ab0c016c015a46734930328f1c25302192aa4945b0bd" Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.223534 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4chpg"] Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.237092 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4chpg"] Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.279132 4775 scope.go:117] "RemoveContainer" containerID="711a82a5ea8d72323dd172e53eaa58caf06be2ca6f1b2822acee6887a437466b" Mar 21 05:34:04 crc kubenswrapper[4775]: E0321 05:34:04.279745 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"711a82a5ea8d72323dd172e53eaa58caf06be2ca6f1b2822acee6887a437466b\": container with ID starting with 711a82a5ea8d72323dd172e53eaa58caf06be2ca6f1b2822acee6887a437466b not found: ID does not exist" containerID="711a82a5ea8d72323dd172e53eaa58caf06be2ca6f1b2822acee6887a437466b" Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.279840 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"711a82a5ea8d72323dd172e53eaa58caf06be2ca6f1b2822acee6887a437466b"} err="failed to get container status \"711a82a5ea8d72323dd172e53eaa58caf06be2ca6f1b2822acee6887a437466b\": rpc error: code = NotFound desc = could not find container \"711a82a5ea8d72323dd172e53eaa58caf06be2ca6f1b2822acee6887a437466b\": container with ID starting with 711a82a5ea8d72323dd172e53eaa58caf06be2ca6f1b2822acee6887a437466b not found: ID does not exist" Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.279901 4775 scope.go:117] "RemoveContainer" containerID="c1908a0d86835431c41047756017337154557cff7b1ff87a8e9a3ad7ba39c029" Mar 21 05:34:04 crc kubenswrapper[4775]: E0321 05:34:04.280441 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1908a0d86835431c41047756017337154557cff7b1ff87a8e9a3ad7ba39c029\": container with ID starting with c1908a0d86835431c41047756017337154557cff7b1ff87a8e9a3ad7ba39c029 not found: ID does not exist" containerID="c1908a0d86835431c41047756017337154557cff7b1ff87a8e9a3ad7ba39c029" Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.280497 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1908a0d86835431c41047756017337154557cff7b1ff87a8e9a3ad7ba39c029"} err="failed to get container status \"c1908a0d86835431c41047756017337154557cff7b1ff87a8e9a3ad7ba39c029\": rpc error: code = NotFound desc = could not find container \"c1908a0d86835431c41047756017337154557cff7b1ff87a8e9a3ad7ba39c029\": container with ID starting with c1908a0d86835431c41047756017337154557cff7b1ff87a8e9a3ad7ba39c029 not found: ID does not exist" Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.280519 4775 scope.go:117] "RemoveContainer" containerID="4237a001ec9e64bac729ab0c016c015a46734930328f1c25302192aa4945b0bd" Mar 21 05:34:04 crc kubenswrapper[4775]: E0321 05:34:04.280897 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4237a001ec9e64bac729ab0c016c015a46734930328f1c25302192aa4945b0bd\": container with ID starting with 4237a001ec9e64bac729ab0c016c015a46734930328f1c25302192aa4945b0bd not found: ID does not exist" containerID="4237a001ec9e64bac729ab0c016c015a46734930328f1c25302192aa4945b0bd" Mar 21 05:34:04 crc kubenswrapper[4775]: I0321 05:34:04.280935 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4237a001ec9e64bac729ab0c016c015a46734930328f1c25302192aa4945b0bd"} err="failed to get container status \"4237a001ec9e64bac729ab0c016c015a46734930328f1c25302192aa4945b0bd\": rpc error: code = NotFound desc = could not find container \"4237a001ec9e64bac729ab0c016c015a46734930328f1c25302192aa4945b0bd\": container with ID starting with 4237a001ec9e64bac729ab0c016c015a46734930328f1c25302192aa4945b0bd not found: ID does not exist" Mar 21 05:34:05 crc kubenswrapper[4775]: I0321 05:34:05.578531 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567854-gxf82" Mar 21 05:34:05 crc kubenswrapper[4775]: I0321 05:34:05.668629 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7kzr\" (UniqueName: \"kubernetes.io/projected/afdb9c39-b2aa-496c-8cfb-917fd0b1cc15-kube-api-access-b7kzr\") pod \"afdb9c39-b2aa-496c-8cfb-917fd0b1cc15\" (UID: \"afdb9c39-b2aa-496c-8cfb-917fd0b1cc15\") " Mar 21 05:34:05 crc kubenswrapper[4775]: I0321 05:34:05.672575 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="501afe19-8db4-4d25-8e06-56d1449b3643" path="/var/lib/kubelet/pods/501afe19-8db4-4d25-8e06-56d1449b3643/volumes" Mar 21 05:34:05 crc kubenswrapper[4775]: I0321 05:34:05.681545 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afdb9c39-b2aa-496c-8cfb-917fd0b1cc15-kube-api-access-b7kzr" (OuterVolumeSpecName: "kube-api-access-b7kzr") pod "afdb9c39-b2aa-496c-8cfb-917fd0b1cc15" (UID: "afdb9c39-b2aa-496c-8cfb-917fd0b1cc15"). InnerVolumeSpecName "kube-api-access-b7kzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:34:05 crc kubenswrapper[4775]: I0321 05:34:05.772197 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7kzr\" (UniqueName: \"kubernetes.io/projected/afdb9c39-b2aa-496c-8cfb-917fd0b1cc15-kube-api-access-b7kzr\") on node \"crc\" DevicePath \"\"" Mar 21 05:34:06 crc kubenswrapper[4775]: I0321 05:34:06.180334 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567854-gxf82" event={"ID":"afdb9c39-b2aa-496c-8cfb-917fd0b1cc15","Type":"ContainerDied","Data":"0d244af90f1f3cc12a466a4d5d229c239003d7b502e8be0d3f5edcd24ca33e5a"} Mar 21 05:34:06 crc kubenswrapper[4775]: I0321 05:34:06.180386 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d244af90f1f3cc12a466a4d5d229c239003d7b502e8be0d3f5edcd24ca33e5a" Mar 21 05:34:06 crc kubenswrapper[4775]: I0321 05:34:06.180460 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567854-gxf82" Mar 21 05:34:06 crc kubenswrapper[4775]: I0321 05:34:06.244087 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567848-6lggf"] Mar 21 05:34:06 crc kubenswrapper[4775]: I0321 05:34:06.255596 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567848-6lggf"] Mar 21 05:34:07 crc kubenswrapper[4775]: I0321 05:34:07.683723 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d6c8144-5186-4a29-a208-b929f47b1695" path="/var/lib/kubelet/pods/5d6c8144-5186-4a29-a208-b929f47b1695/volumes" Mar 21 05:34:48 crc kubenswrapper[4775]: I0321 05:34:48.042931 4775 scope.go:117] "RemoveContainer" containerID="b0bde1a2f3bc4718dd2d5ab50594a32268b2724339820463f90a790bd3794b31" Mar 21 05:35:42 crc kubenswrapper[4775]: I0321 05:35:42.461170 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-frvc6"] Mar 21 05:35:42 crc kubenswrapper[4775]: E0321 05:35:42.462415 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afdb9c39-b2aa-496c-8cfb-917fd0b1cc15" containerName="oc" Mar 21 05:35:42 crc kubenswrapper[4775]: I0321 05:35:42.462430 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="afdb9c39-b2aa-496c-8cfb-917fd0b1cc15" containerName="oc" Mar 21 05:35:42 crc kubenswrapper[4775]: E0321 05:35:42.462444 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501afe19-8db4-4d25-8e06-56d1449b3643" containerName="extract-utilities" Mar 21 05:35:42 crc kubenswrapper[4775]: I0321 05:35:42.462451 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="501afe19-8db4-4d25-8e06-56d1449b3643" containerName="extract-utilities" Mar 21 05:35:42 crc kubenswrapper[4775]: E0321 05:35:42.462474 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501afe19-8db4-4d25-8e06-56d1449b3643" containerName="extract-content" Mar 21 05:35:42 crc kubenswrapper[4775]: I0321 05:35:42.462480 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="501afe19-8db4-4d25-8e06-56d1449b3643" containerName="extract-content" Mar 21 05:35:42 crc kubenswrapper[4775]: E0321 05:35:42.462505 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501afe19-8db4-4d25-8e06-56d1449b3643" containerName="registry-server" Mar 21 05:35:42 crc kubenswrapper[4775]: I0321 05:35:42.462511 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="501afe19-8db4-4d25-8e06-56d1449b3643" containerName="registry-server" Mar 21 05:35:42 crc kubenswrapper[4775]: I0321 05:35:42.462697 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="afdb9c39-b2aa-496c-8cfb-917fd0b1cc15" containerName="oc" Mar 21 05:35:42 crc kubenswrapper[4775]: I0321 05:35:42.462713 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="501afe19-8db4-4d25-8e06-56d1449b3643" containerName="registry-server" Mar 21 05:35:42 crc kubenswrapper[4775]: I0321 05:35:42.464176 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-frvc6" Mar 21 05:35:42 crc kubenswrapper[4775]: I0321 05:35:42.478895 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-frvc6"] Mar 21 05:35:42 crc kubenswrapper[4775]: I0321 05:35:42.594702 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4-catalog-content\") pod \"community-operators-frvc6\" (UID: \"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4\") " pod="openshift-marketplace/community-operators-frvc6" Mar 21 05:35:42 crc kubenswrapper[4775]: I0321 05:35:42.594784 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4-utilities\") pod \"community-operators-frvc6\" (UID: \"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4\") " pod="openshift-marketplace/community-operators-frvc6" Mar 21 05:35:42 crc kubenswrapper[4775]: I0321 05:35:42.594864 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spsl5\" (UniqueName: \"kubernetes.io/projected/7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4-kube-api-access-spsl5\") pod \"community-operators-frvc6\" (UID: \"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4\") " pod="openshift-marketplace/community-operators-frvc6" Mar 21 05:35:42 crc kubenswrapper[4775]: I0321 05:35:42.697297 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4-catalog-content\") pod \"community-operators-frvc6\" (UID: \"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4\") " pod="openshift-marketplace/community-operators-frvc6" Mar 21 05:35:42 crc kubenswrapper[4775]: I0321 05:35:42.697362 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4-utilities\") pod \"community-operators-frvc6\" (UID: \"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4\") " pod="openshift-marketplace/community-operators-frvc6" Mar 21 05:35:42 crc kubenswrapper[4775]: I0321 05:35:42.697423 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spsl5\" (UniqueName: \"kubernetes.io/projected/7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4-kube-api-access-spsl5\") pod \"community-operators-frvc6\" (UID: \"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4\") " pod="openshift-marketplace/community-operators-frvc6" Mar 21 05:35:42 crc kubenswrapper[4775]: I0321 05:35:42.698484 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4-catalog-content\") pod \"community-operators-frvc6\" (UID: \"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4\") " pod="openshift-marketplace/community-operators-frvc6" Mar 21 05:35:42 crc kubenswrapper[4775]: I0321 05:35:42.699141 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4-utilities\") pod \"community-operators-frvc6\" (UID: \"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4\") " pod="openshift-marketplace/community-operators-frvc6" Mar 21 05:35:42 crc kubenswrapper[4775]: I0321 05:35:42.720785 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spsl5\" (UniqueName: \"kubernetes.io/projected/7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4-kube-api-access-spsl5\") pod \"community-operators-frvc6\" (UID: \"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4\") " pod="openshift-marketplace/community-operators-frvc6" Mar 21 05:35:42 crc kubenswrapper[4775]: I0321 05:35:42.794421 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-frvc6" Mar 21 05:35:43 crc kubenswrapper[4775]: I0321 05:35:43.438241 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-frvc6"] Mar 21 05:35:44 crc kubenswrapper[4775]: I0321 05:35:44.047287 4775 generic.go:334] "Generic (PLEG): container finished" podID="7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4" containerID="db650e0f182ffb5e56cbba3ab5adf162afa3a564247002767dee7622fc9639de" exitCode=0 Mar 21 05:35:44 crc kubenswrapper[4775]: I0321 05:35:44.047399 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-frvc6" event={"ID":"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4","Type":"ContainerDied","Data":"db650e0f182ffb5e56cbba3ab5adf162afa3a564247002767dee7622fc9639de"} Mar 21 05:35:44 crc kubenswrapper[4775]: I0321 05:35:44.047704 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-frvc6" event={"ID":"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4","Type":"ContainerStarted","Data":"8ac4aab02208f399e31bb5fece0a658d82b8219e89b0629a802dfd4e78ef1fe1"} Mar 21 05:35:46 crc kubenswrapper[4775]: I0321 05:35:46.237797 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xqxgq"] Mar 21 05:35:46 crc kubenswrapper[4775]: I0321 05:35:46.241359 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xqxgq" Mar 21 05:35:46 crc kubenswrapper[4775]: I0321 05:35:46.269029 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xqxgq"] Mar 21 05:35:46 crc kubenswrapper[4775]: I0321 05:35:46.328699 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd-utilities\") pod \"redhat-marketplace-xqxgq\" (UID: \"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd\") " pod="openshift-marketplace/redhat-marketplace-xqxgq" Mar 21 05:35:46 crc kubenswrapper[4775]: I0321 05:35:46.328848 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd-catalog-content\") pod \"redhat-marketplace-xqxgq\" (UID: \"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd\") " pod="openshift-marketplace/redhat-marketplace-xqxgq" Mar 21 05:35:46 crc kubenswrapper[4775]: I0321 05:35:46.328960 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtkjr\" (UniqueName: \"kubernetes.io/projected/d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd-kube-api-access-xtkjr\") pod \"redhat-marketplace-xqxgq\" (UID: \"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd\") " pod="openshift-marketplace/redhat-marketplace-xqxgq" Mar 21 05:35:46 crc kubenswrapper[4775]: I0321 05:35:46.431669 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd-catalog-content\") pod \"redhat-marketplace-xqxgq\" (UID: \"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd\") " pod="openshift-marketplace/redhat-marketplace-xqxgq" Mar 21 05:35:46 crc kubenswrapper[4775]: I0321 05:35:46.431842 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtkjr\" (UniqueName: \"kubernetes.io/projected/d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd-kube-api-access-xtkjr\") pod \"redhat-marketplace-xqxgq\" (UID: \"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd\") " pod="openshift-marketplace/redhat-marketplace-xqxgq" Mar 21 05:35:46 crc kubenswrapper[4775]: I0321 05:35:46.431955 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd-utilities\") pod \"redhat-marketplace-xqxgq\" (UID: \"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd\") " pod="openshift-marketplace/redhat-marketplace-xqxgq" Mar 21 05:35:46 crc kubenswrapper[4775]: I0321 05:35:46.432477 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd-utilities\") pod \"redhat-marketplace-xqxgq\" (UID: \"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd\") " pod="openshift-marketplace/redhat-marketplace-xqxgq" Mar 21 05:35:46 crc kubenswrapper[4775]: I0321 05:35:46.432738 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd-catalog-content\") pod \"redhat-marketplace-xqxgq\" (UID: \"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd\") " pod="openshift-marketplace/redhat-marketplace-xqxgq" Mar 21 05:35:46 crc kubenswrapper[4775]: I0321 05:35:46.455569 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtkjr\" (UniqueName: \"kubernetes.io/projected/d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd-kube-api-access-xtkjr\") pod \"redhat-marketplace-xqxgq\" (UID: \"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd\") " pod="openshift-marketplace/redhat-marketplace-xqxgq" Mar 21 05:35:46 crc kubenswrapper[4775]: I0321 05:35:46.616416 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xqxgq" Mar 21 05:35:46 crc kubenswrapper[4775]: I0321 05:35:46.770357 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-frvc6" event={"ID":"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4","Type":"ContainerStarted","Data":"8df7790e4cf98b44556ff5c488018f03fe2a353bc67457fc94935e62f20efe2f"} Mar 21 05:35:47 crc kubenswrapper[4775]: I0321 05:35:47.162469 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xqxgq"] Mar 21 05:35:47 crc kubenswrapper[4775]: I0321 05:35:47.782175 4775 generic.go:334] "Generic (PLEG): container finished" podID="d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd" containerID="7555adb7dfe7a1ef65a02bb7b67511ce5cb7651c9d7d951e7c2e95e2b9054af6" exitCode=0 Mar 21 05:35:47 crc kubenswrapper[4775]: I0321 05:35:47.782242 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xqxgq" event={"ID":"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd","Type":"ContainerDied","Data":"7555adb7dfe7a1ef65a02bb7b67511ce5cb7651c9d7d951e7c2e95e2b9054af6"} Mar 21 05:35:47 crc kubenswrapper[4775]: I0321 05:35:47.783658 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xqxgq" event={"ID":"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd","Type":"ContainerStarted","Data":"2e22d3f3fe30d01cb59c63a65374d6f6523a902e99ea14cb36acbf8af271d8b6"} Mar 21 05:35:48 crc kubenswrapper[4775]: I0321 05:35:48.794943 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xqxgq" event={"ID":"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd","Type":"ContainerStarted","Data":"8305d05030b8b6f9feaa1265bdc85f418d6ed22420c9ec22443e3decb737048f"} Mar 21 05:35:48 crc kubenswrapper[4775]: I0321 05:35:48.798666 4775 generic.go:334] "Generic (PLEG): container finished" podID="7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4" containerID="8df7790e4cf98b44556ff5c488018f03fe2a353bc67457fc94935e62f20efe2f" exitCode=0 Mar 21 05:35:48 crc kubenswrapper[4775]: I0321 05:35:48.798736 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-frvc6" event={"ID":"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4","Type":"ContainerDied","Data":"8df7790e4cf98b44556ff5c488018f03fe2a353bc67457fc94935e62f20efe2f"} Mar 21 05:35:49 crc kubenswrapper[4775]: I0321 05:35:49.816049 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-frvc6" event={"ID":"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4","Type":"ContainerStarted","Data":"709878fa6db14a1a4c0041409907d252d31e1ef59aa788ccad3dcfcb237f419c"} Mar 21 05:35:49 crc kubenswrapper[4775]: I0321 05:35:49.820237 4775 generic.go:334] "Generic (PLEG): container finished" podID="d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd" containerID="8305d05030b8b6f9feaa1265bdc85f418d6ed22420c9ec22443e3decb737048f" exitCode=0 Mar 21 05:35:49 crc kubenswrapper[4775]: I0321 05:35:49.820413 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xqxgq" event={"ID":"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd","Type":"ContainerDied","Data":"8305d05030b8b6f9feaa1265bdc85f418d6ed22420c9ec22443e3decb737048f"} Mar 21 05:35:49 crc kubenswrapper[4775]: I0321 05:35:49.847962 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-frvc6" podStartSLOduration=2.658405369 podStartE2EDuration="7.847939603s" podCreationTimestamp="2026-03-21 05:35:42 +0000 UTC" firstStartedPulling="2026-03-21 05:35:44.050009914 +0000 UTC m=+2897.026473538" lastFinishedPulling="2026-03-21 05:35:49.239544148 +0000 UTC m=+2902.216007772" observedRunningTime="2026-03-21 05:35:49.840495118 +0000 UTC m=+2902.816958742" watchObservedRunningTime="2026-03-21 05:35:49.847939603 +0000 UTC m=+2902.824403237" Mar 21 05:35:50 crc kubenswrapper[4775]: I0321 05:35:50.849507 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xqxgq" event={"ID":"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd","Type":"ContainerStarted","Data":"30e789e403772d7ffce3812d73fb279cd48892d4950ddba21d34141ef557b9c2"} Mar 21 05:35:50 crc kubenswrapper[4775]: I0321 05:35:50.873410 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xqxgq" podStartSLOduration=2.427559941 podStartE2EDuration="4.87338716s" podCreationTimestamp="2026-03-21 05:35:46 +0000 UTC" firstStartedPulling="2026-03-21 05:35:47.78496959 +0000 UTC m=+2900.761433214" lastFinishedPulling="2026-03-21 05:35:50.230796809 +0000 UTC m=+2903.207260433" observedRunningTime="2026-03-21 05:35:50.872529786 +0000 UTC m=+2903.848993420" watchObservedRunningTime="2026-03-21 05:35:50.87338716 +0000 UTC m=+2903.849850784" Mar 21 05:35:52 crc kubenswrapper[4775]: I0321 05:35:52.794920 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-frvc6" Mar 21 05:35:52 crc kubenswrapper[4775]: I0321 05:35:52.796341 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-frvc6" Mar 21 05:35:53 crc kubenswrapper[4775]: I0321 05:35:53.849683 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-frvc6" podUID="7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4" containerName="registry-server" probeResult="failure" output=< Mar 21 05:35:53 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Mar 21 05:35:53 crc kubenswrapper[4775]: > Mar 21 05:35:56 crc kubenswrapper[4775]: I0321 05:35:56.616907 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xqxgq" Mar 21 05:35:56 crc kubenswrapper[4775]: I0321 05:35:56.617339 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xqxgq" Mar 21 05:35:56 crc kubenswrapper[4775]: I0321 05:35:56.670245 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xqxgq" Mar 21 05:35:56 crc kubenswrapper[4775]: I0321 05:35:56.955288 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xqxgq" Mar 21 05:35:57 crc kubenswrapper[4775]: I0321 05:35:57.011287 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xqxgq"] Mar 21 05:35:58 crc kubenswrapper[4775]: I0321 05:35:58.940328 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xqxgq" podUID="d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd" containerName="registry-server" containerID="cri-o://30e789e403772d7ffce3812d73fb279cd48892d4950ddba21d34141ef557b9c2" gracePeriod=2 Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:35:59.572300 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xqxgq" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:35:59.745847 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd-utilities\") pod \"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd\" (UID: \"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd\") " Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:35:59.746012 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtkjr\" (UniqueName: \"kubernetes.io/projected/d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd-kube-api-access-xtkjr\") pod \"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd\" (UID: \"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd\") " Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:35:59.746165 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd-catalog-content\") pod \"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd\" (UID: \"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd\") " Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:35:59.747065 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd-utilities" (OuterVolumeSpecName: "utilities") pod "d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd" (UID: "d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:35:59.753099 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd-kube-api-access-xtkjr" (OuterVolumeSpecName: "kube-api-access-xtkjr") pod "d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd" (UID: "d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd"). InnerVolumeSpecName "kube-api-access-xtkjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:35:59.772571 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd" (UID: "d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:35:59.849349 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:35:59.849391 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtkjr\" (UniqueName: \"kubernetes.io/projected/d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd-kube-api-access-xtkjr\") on node \"crc\" DevicePath \"\"" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:35:59.849401 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:35:59.951052 4775 generic.go:334] "Generic (PLEG): container finished" podID="d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd" containerID="30e789e403772d7ffce3812d73fb279cd48892d4950ddba21d34141ef557b9c2" exitCode=0 Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:35:59.951096 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xqxgq" event={"ID":"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd","Type":"ContainerDied","Data":"30e789e403772d7ffce3812d73fb279cd48892d4950ddba21d34141ef557b9c2"} Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:35:59.951163 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xqxgq" event={"ID":"d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd","Type":"ContainerDied","Data":"2e22d3f3fe30d01cb59c63a65374d6f6523a902e99ea14cb36acbf8af271d8b6"} Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:35:59.951184 4775 scope.go:117] "RemoveContainer" containerID="30e789e403772d7ffce3812d73fb279cd48892d4950ddba21d34141ef557b9c2" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:35:59.951201 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xqxgq" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:35:59.984415 4775 scope.go:117] "RemoveContainer" containerID="8305d05030b8b6f9feaa1265bdc85f418d6ed22420c9ec22443e3decb737048f" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:35:59.993936 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xqxgq"] Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.003592 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xqxgq"] Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.009334 4775 scope.go:117] "RemoveContainer" containerID="7555adb7dfe7a1ef65a02bb7b67511ce5cb7651c9d7d951e7c2e95e2b9054af6" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.062275 4775 scope.go:117] "RemoveContainer" containerID="30e789e403772d7ffce3812d73fb279cd48892d4950ddba21d34141ef557b9c2" Mar 21 05:36:01 crc kubenswrapper[4775]: E0321 05:36:00.062887 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30e789e403772d7ffce3812d73fb279cd48892d4950ddba21d34141ef557b9c2\": container with ID starting with 30e789e403772d7ffce3812d73fb279cd48892d4950ddba21d34141ef557b9c2 not found: ID does not exist" containerID="30e789e403772d7ffce3812d73fb279cd48892d4950ddba21d34141ef557b9c2" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.062925 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30e789e403772d7ffce3812d73fb279cd48892d4950ddba21d34141ef557b9c2"} err="failed to get container status \"30e789e403772d7ffce3812d73fb279cd48892d4950ddba21d34141ef557b9c2\": rpc error: code = NotFound desc = could not find container \"30e789e403772d7ffce3812d73fb279cd48892d4950ddba21d34141ef557b9c2\": container with ID starting with 30e789e403772d7ffce3812d73fb279cd48892d4950ddba21d34141ef557b9c2 not found: ID does not exist" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.062947 4775 scope.go:117] "RemoveContainer" containerID="8305d05030b8b6f9feaa1265bdc85f418d6ed22420c9ec22443e3decb737048f" Mar 21 05:36:01 crc kubenswrapper[4775]: E0321 05:36:00.063436 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8305d05030b8b6f9feaa1265bdc85f418d6ed22420c9ec22443e3decb737048f\": container with ID starting with 8305d05030b8b6f9feaa1265bdc85f418d6ed22420c9ec22443e3decb737048f not found: ID does not exist" containerID="8305d05030b8b6f9feaa1265bdc85f418d6ed22420c9ec22443e3decb737048f" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.063511 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8305d05030b8b6f9feaa1265bdc85f418d6ed22420c9ec22443e3decb737048f"} err="failed to get container status \"8305d05030b8b6f9feaa1265bdc85f418d6ed22420c9ec22443e3decb737048f\": rpc error: code = NotFound desc = could not find container \"8305d05030b8b6f9feaa1265bdc85f418d6ed22420c9ec22443e3decb737048f\": container with ID starting with 8305d05030b8b6f9feaa1265bdc85f418d6ed22420c9ec22443e3decb737048f not found: ID does not exist" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.063541 4775 scope.go:117] "RemoveContainer" containerID="7555adb7dfe7a1ef65a02bb7b67511ce5cb7651c9d7d951e7c2e95e2b9054af6" Mar 21 05:36:01 crc kubenswrapper[4775]: E0321 05:36:00.064055 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7555adb7dfe7a1ef65a02bb7b67511ce5cb7651c9d7d951e7c2e95e2b9054af6\": container with ID starting with 7555adb7dfe7a1ef65a02bb7b67511ce5cb7651c9d7d951e7c2e95e2b9054af6 not found: ID does not exist" containerID="7555adb7dfe7a1ef65a02bb7b67511ce5cb7651c9d7d951e7c2e95e2b9054af6" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.064083 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7555adb7dfe7a1ef65a02bb7b67511ce5cb7651c9d7d951e7c2e95e2b9054af6"} err="failed to get container status \"7555adb7dfe7a1ef65a02bb7b67511ce5cb7651c9d7d951e7c2e95e2b9054af6\": rpc error: code = NotFound desc = could not find container \"7555adb7dfe7a1ef65a02bb7b67511ce5cb7651c9d7d951e7c2e95e2b9054af6\": container with ID starting with 7555adb7dfe7a1ef65a02bb7b67511ce5cb7651c9d7d951e7c2e95e2b9054af6 not found: ID does not exist" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.155036 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567856-vjs55"] Mar 21 05:36:01 crc kubenswrapper[4775]: E0321 05:36:00.155740 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd" containerName="extract-utilities" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.155761 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd" containerName="extract-utilities" Mar 21 05:36:01 crc kubenswrapper[4775]: E0321 05:36:00.155775 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd" containerName="extract-content" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.155804 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd" containerName="extract-content" Mar 21 05:36:01 crc kubenswrapper[4775]: E0321 05:36:00.155845 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd" containerName="registry-server" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.155853 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd" containerName="registry-server" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.156097 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd" containerName="registry-server" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.156883 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567856-vjs55" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.159778 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.160075 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.160438 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.170935 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567856-vjs55"] Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.256648 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtwxb\" (UniqueName: \"kubernetes.io/projected/d0b49c68-9fa7-444c-b3cd-4bef5eff4765-kube-api-access-mtwxb\") pod \"auto-csr-approver-29567856-vjs55\" (UID: \"d0b49c68-9fa7-444c-b3cd-4bef5eff4765\") " pod="openshift-infra/auto-csr-approver-29567856-vjs55" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.359187 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtwxb\" (UniqueName: \"kubernetes.io/projected/d0b49c68-9fa7-444c-b3cd-4bef5eff4765-kube-api-access-mtwxb\") pod \"auto-csr-approver-29567856-vjs55\" (UID: \"d0b49c68-9fa7-444c-b3cd-4bef5eff4765\") " pod="openshift-infra/auto-csr-approver-29567856-vjs55" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.395620 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtwxb\" (UniqueName: \"kubernetes.io/projected/d0b49c68-9fa7-444c-b3cd-4bef5eff4765-kube-api-access-mtwxb\") pod \"auto-csr-approver-29567856-vjs55\" (UID: \"d0b49c68-9fa7-444c-b3cd-4bef5eff4765\") " pod="openshift-infra/auto-csr-approver-29567856-vjs55" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:00.479323 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567856-vjs55" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:01.673955 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd" path="/var/lib/kubelet/pods/d1cf9da2-cb82-4e41-b9e6-5830ab22e5bd/volumes" Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:01.714996 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567856-vjs55"] Mar 21 05:36:01 crc kubenswrapper[4775]: W0321 05:36:01.726349 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0b49c68_9fa7_444c_b3cd_4bef5eff4765.slice/crio-5eb1c74e4a6841837e571124111b7f03c3a51fcbc60071ac45babd10071eab92 WatchSource:0}: Error finding container 5eb1c74e4a6841837e571124111b7f03c3a51fcbc60071ac45babd10071eab92: Status 404 returned error can't find the container with id 5eb1c74e4a6841837e571124111b7f03c3a51fcbc60071ac45babd10071eab92 Mar 21 05:36:01 crc kubenswrapper[4775]: I0321 05:36:01.971709 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567856-vjs55" event={"ID":"d0b49c68-9fa7-444c-b3cd-4bef5eff4765","Type":"ContainerStarted","Data":"5eb1c74e4a6841837e571124111b7f03c3a51fcbc60071ac45babd10071eab92"} Mar 21 05:36:02 crc kubenswrapper[4775]: I0321 05:36:02.861998 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-frvc6" Mar 21 05:36:02 crc kubenswrapper[4775]: I0321 05:36:02.993547 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-frvc6" Mar 21 05:36:03 crc kubenswrapper[4775]: I0321 05:36:03.314746 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-frvc6"] Mar 21 05:36:04 crc kubenswrapper[4775]: I0321 05:36:04.008887 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-frvc6" podUID="7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4" containerName="registry-server" containerID="cri-o://709878fa6db14a1a4c0041409907d252d31e1ef59aa788ccad3dcfcb237f419c" gracePeriod=2 Mar 21 05:36:04 crc kubenswrapper[4775]: I0321 05:36:04.619320 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-frvc6" Mar 21 05:36:04 crc kubenswrapper[4775]: I0321 05:36:04.678071 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4-utilities\") pod \"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4\" (UID: \"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4\") " Mar 21 05:36:04 crc kubenswrapper[4775]: I0321 05:36:04.678314 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spsl5\" (UniqueName: \"kubernetes.io/projected/7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4-kube-api-access-spsl5\") pod \"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4\" (UID: \"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4\") " Mar 21 05:36:04 crc kubenswrapper[4775]: I0321 05:36:04.678488 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4-catalog-content\") pod \"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4\" (UID: \"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4\") " Mar 21 05:36:04 crc kubenswrapper[4775]: I0321 05:36:04.679671 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4-utilities" (OuterVolumeSpecName: "utilities") pod "7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4" (UID: "7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:36:04 crc kubenswrapper[4775]: I0321 05:36:04.686844 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4-kube-api-access-spsl5" (OuterVolumeSpecName: "kube-api-access-spsl5") pod "7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4" (UID: "7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4"). InnerVolumeSpecName "kube-api-access-spsl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:36:04 crc kubenswrapper[4775]: I0321 05:36:04.745941 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4" (UID: "7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:36:04 crc kubenswrapper[4775]: I0321 05:36:04.781409 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:36:04 crc kubenswrapper[4775]: I0321 05:36:04.781453 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:36:04 crc kubenswrapper[4775]: I0321 05:36:04.781471 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spsl5\" (UniqueName: \"kubernetes.io/projected/7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4-kube-api-access-spsl5\") on node \"crc\" DevicePath \"\"" Mar 21 05:36:05 crc kubenswrapper[4775]: I0321 05:36:05.025385 4775 generic.go:334] "Generic (PLEG): container finished" podID="7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4" containerID="709878fa6db14a1a4c0041409907d252d31e1ef59aa788ccad3dcfcb237f419c" exitCode=0 Mar 21 05:36:05 crc kubenswrapper[4775]: I0321 05:36:05.025468 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-frvc6" Mar 21 05:36:05 crc kubenswrapper[4775]: I0321 05:36:05.025482 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-frvc6" event={"ID":"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4","Type":"ContainerDied","Data":"709878fa6db14a1a4c0041409907d252d31e1ef59aa788ccad3dcfcb237f419c"} Mar 21 05:36:05 crc kubenswrapper[4775]: I0321 05:36:05.025702 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-frvc6" event={"ID":"7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4","Type":"ContainerDied","Data":"8ac4aab02208f399e31bb5fece0a658d82b8219e89b0629a802dfd4e78ef1fe1"} Mar 21 05:36:05 crc kubenswrapper[4775]: I0321 05:36:05.025729 4775 scope.go:117] "RemoveContainer" containerID="709878fa6db14a1a4c0041409907d252d31e1ef59aa788ccad3dcfcb237f419c" Mar 21 05:36:05 crc kubenswrapper[4775]: I0321 05:36:05.030392 4775 generic.go:334] "Generic (PLEG): container finished" podID="d0b49c68-9fa7-444c-b3cd-4bef5eff4765" containerID="fe0d6a170f19e7078dcaeb489e7d9b4bba7aef3e240f5e59424f9dea5b0d382f" exitCode=0 Mar 21 05:36:05 crc kubenswrapper[4775]: I0321 05:36:05.030454 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567856-vjs55" event={"ID":"d0b49c68-9fa7-444c-b3cd-4bef5eff4765","Type":"ContainerDied","Data":"fe0d6a170f19e7078dcaeb489e7d9b4bba7aef3e240f5e59424f9dea5b0d382f"} Mar 21 05:36:05 crc kubenswrapper[4775]: I0321 05:36:05.053869 4775 scope.go:117] "RemoveContainer" containerID="8df7790e4cf98b44556ff5c488018f03fe2a353bc67457fc94935e62f20efe2f" Mar 21 05:36:05 crc kubenswrapper[4775]: I0321 05:36:05.072255 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-frvc6"] Mar 21 05:36:05 crc kubenswrapper[4775]: I0321 05:36:05.080237 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-frvc6"] Mar 21 05:36:05 crc kubenswrapper[4775]: I0321 05:36:05.109233 4775 scope.go:117] "RemoveContainer" containerID="db650e0f182ffb5e56cbba3ab5adf162afa3a564247002767dee7622fc9639de" Mar 21 05:36:05 crc kubenswrapper[4775]: I0321 05:36:05.134540 4775 scope.go:117] "RemoveContainer" containerID="709878fa6db14a1a4c0041409907d252d31e1ef59aa788ccad3dcfcb237f419c" Mar 21 05:36:05 crc kubenswrapper[4775]: E0321 05:36:05.135091 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"709878fa6db14a1a4c0041409907d252d31e1ef59aa788ccad3dcfcb237f419c\": container with ID starting with 709878fa6db14a1a4c0041409907d252d31e1ef59aa788ccad3dcfcb237f419c not found: ID does not exist" containerID="709878fa6db14a1a4c0041409907d252d31e1ef59aa788ccad3dcfcb237f419c" Mar 21 05:36:05 crc kubenswrapper[4775]: I0321 05:36:05.135232 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"709878fa6db14a1a4c0041409907d252d31e1ef59aa788ccad3dcfcb237f419c"} err="failed to get container status \"709878fa6db14a1a4c0041409907d252d31e1ef59aa788ccad3dcfcb237f419c\": rpc error: code = NotFound desc = could not find container \"709878fa6db14a1a4c0041409907d252d31e1ef59aa788ccad3dcfcb237f419c\": container with ID starting with 709878fa6db14a1a4c0041409907d252d31e1ef59aa788ccad3dcfcb237f419c not found: ID does not exist" Mar 21 05:36:05 crc kubenswrapper[4775]: I0321 05:36:05.135272 4775 scope.go:117] "RemoveContainer" containerID="8df7790e4cf98b44556ff5c488018f03fe2a353bc67457fc94935e62f20efe2f" Mar 21 05:36:05 crc kubenswrapper[4775]: E0321 05:36:05.135722 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8df7790e4cf98b44556ff5c488018f03fe2a353bc67457fc94935e62f20efe2f\": container with ID starting with 8df7790e4cf98b44556ff5c488018f03fe2a353bc67457fc94935e62f20efe2f not found: ID does not exist" containerID="8df7790e4cf98b44556ff5c488018f03fe2a353bc67457fc94935e62f20efe2f" Mar 21 05:36:05 crc kubenswrapper[4775]: I0321 05:36:05.135802 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8df7790e4cf98b44556ff5c488018f03fe2a353bc67457fc94935e62f20efe2f"} err="failed to get container status \"8df7790e4cf98b44556ff5c488018f03fe2a353bc67457fc94935e62f20efe2f\": rpc error: code = NotFound desc = could not find container \"8df7790e4cf98b44556ff5c488018f03fe2a353bc67457fc94935e62f20efe2f\": container with ID starting with 8df7790e4cf98b44556ff5c488018f03fe2a353bc67457fc94935e62f20efe2f not found: ID does not exist" Mar 21 05:36:05 crc kubenswrapper[4775]: I0321 05:36:05.135832 4775 scope.go:117] "RemoveContainer" containerID="db650e0f182ffb5e56cbba3ab5adf162afa3a564247002767dee7622fc9639de" Mar 21 05:36:05 crc kubenswrapper[4775]: E0321 05:36:05.136179 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db650e0f182ffb5e56cbba3ab5adf162afa3a564247002767dee7622fc9639de\": container with ID starting with db650e0f182ffb5e56cbba3ab5adf162afa3a564247002767dee7622fc9639de not found: ID does not exist" containerID="db650e0f182ffb5e56cbba3ab5adf162afa3a564247002767dee7622fc9639de" Mar 21 05:36:05 crc kubenswrapper[4775]: I0321 05:36:05.136207 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db650e0f182ffb5e56cbba3ab5adf162afa3a564247002767dee7622fc9639de"} err="failed to get container status \"db650e0f182ffb5e56cbba3ab5adf162afa3a564247002767dee7622fc9639de\": rpc error: code = NotFound desc = could not find container \"db650e0f182ffb5e56cbba3ab5adf162afa3a564247002767dee7622fc9639de\": container with ID starting with db650e0f182ffb5e56cbba3ab5adf162afa3a564247002767dee7622fc9639de not found: ID does not exist" Mar 21 05:36:05 crc kubenswrapper[4775]: I0321 05:36:05.673174 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4" path="/var/lib/kubelet/pods/7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4/volumes" Mar 21 05:36:06 crc kubenswrapper[4775]: I0321 05:36:06.425233 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567856-vjs55" Mar 21 05:36:06 crc kubenswrapper[4775]: I0321 05:36:06.524483 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtwxb\" (UniqueName: \"kubernetes.io/projected/d0b49c68-9fa7-444c-b3cd-4bef5eff4765-kube-api-access-mtwxb\") pod \"d0b49c68-9fa7-444c-b3cd-4bef5eff4765\" (UID: \"d0b49c68-9fa7-444c-b3cd-4bef5eff4765\") " Mar 21 05:36:06 crc kubenswrapper[4775]: I0321 05:36:06.532726 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0b49c68-9fa7-444c-b3cd-4bef5eff4765-kube-api-access-mtwxb" (OuterVolumeSpecName: "kube-api-access-mtwxb") pod "d0b49c68-9fa7-444c-b3cd-4bef5eff4765" (UID: "d0b49c68-9fa7-444c-b3cd-4bef5eff4765"). InnerVolumeSpecName "kube-api-access-mtwxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:36:06 crc kubenswrapper[4775]: I0321 05:36:06.627237 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtwxb\" (UniqueName: \"kubernetes.io/projected/d0b49c68-9fa7-444c-b3cd-4bef5eff4765-kube-api-access-mtwxb\") on node \"crc\" DevicePath \"\"" Mar 21 05:36:07 crc kubenswrapper[4775]: I0321 05:36:07.054407 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567856-vjs55" event={"ID":"d0b49c68-9fa7-444c-b3cd-4bef5eff4765","Type":"ContainerDied","Data":"5eb1c74e4a6841837e571124111b7f03c3a51fcbc60071ac45babd10071eab92"} Mar 21 05:36:07 crc kubenswrapper[4775]: I0321 05:36:07.054459 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5eb1c74e4a6841837e571124111b7f03c3a51fcbc60071ac45babd10071eab92" Mar 21 05:36:07 crc kubenswrapper[4775]: I0321 05:36:07.054769 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567856-vjs55" Mar 21 05:36:07 crc kubenswrapper[4775]: I0321 05:36:07.507717 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567850-gfbv5"] Mar 21 05:36:07 crc kubenswrapper[4775]: I0321 05:36:07.519963 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567850-gfbv5"] Mar 21 05:36:07 crc kubenswrapper[4775]: I0321 05:36:07.671938 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2616ee6e-f4b6-4258-b738-95ad6aab1644" path="/var/lib/kubelet/pods/2616ee6e-f4b6-4258-b738-95ad6aab1644/volumes" Mar 21 05:36:32 crc kubenswrapper[4775]: I0321 05:36:32.482713 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:36:32 crc kubenswrapper[4775]: I0321 05:36:32.483774 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:36:48 crc kubenswrapper[4775]: I0321 05:36:48.171848 4775 scope.go:117] "RemoveContainer" containerID="2262a4a2de084f40220b6b85086a44442e4d874dfee1c84bf3dcc589fdc32198" Mar 21 05:37:02 crc kubenswrapper[4775]: I0321 05:37:02.482373 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:37:02 crc kubenswrapper[4775]: I0321 05:37:02.483020 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:37:32 crc kubenswrapper[4775]: I0321 05:37:32.482584 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:37:32 crc kubenswrapper[4775]: I0321 05:37:32.483105 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:37:32 crc kubenswrapper[4775]: I0321 05:37:32.483168 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 05:37:32 crc kubenswrapper[4775]: I0321 05:37:32.483908 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9"} pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 05:37:32 crc kubenswrapper[4775]: I0321 05:37:32.483952 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" containerID="cri-o://f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" gracePeriod=600 Mar 21 05:37:32 crc kubenswrapper[4775]: E0321 05:37:32.631108 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:37:33 crc kubenswrapper[4775]: I0321 05:37:33.451288 4775 generic.go:334] "Generic (PLEG): container finished" podID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" exitCode=0 Mar 21 05:37:33 crc kubenswrapper[4775]: I0321 05:37:33.451397 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerDied","Data":"f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9"} Mar 21 05:37:33 crc kubenswrapper[4775]: I0321 05:37:33.451897 4775 scope.go:117] "RemoveContainer" containerID="ea40506310b7606d65a9c28447427b7d1714319295a97c7d0f6f7a3b4b8251ee" Mar 21 05:37:33 crc kubenswrapper[4775]: I0321 05:37:33.452843 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:37:33 crc kubenswrapper[4775]: E0321 05:37:33.453166 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:37:44 crc kubenswrapper[4775]: I0321 05:37:44.661477 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:37:44 crc kubenswrapper[4775]: E0321 05:37:44.662603 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:37:57 crc kubenswrapper[4775]: I0321 05:37:57.667767 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:37:57 crc kubenswrapper[4775]: E0321 05:37:57.668690 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:38:00 crc kubenswrapper[4775]: I0321 05:38:00.146668 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567858-n842g"] Mar 21 05:38:00 crc kubenswrapper[4775]: E0321 05:38:00.147485 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4" containerName="extract-utilities" Mar 21 05:38:00 crc kubenswrapper[4775]: I0321 05:38:00.147502 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4" containerName="extract-utilities" Mar 21 05:38:00 crc kubenswrapper[4775]: E0321 05:38:00.147522 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0b49c68-9fa7-444c-b3cd-4bef5eff4765" containerName="oc" Mar 21 05:38:00 crc kubenswrapper[4775]: I0321 05:38:00.147529 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0b49c68-9fa7-444c-b3cd-4bef5eff4765" containerName="oc" Mar 21 05:38:00 crc kubenswrapper[4775]: E0321 05:38:00.147542 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4" containerName="extract-content" Mar 21 05:38:00 crc kubenswrapper[4775]: I0321 05:38:00.147549 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4" containerName="extract-content" Mar 21 05:38:00 crc kubenswrapper[4775]: E0321 05:38:00.147561 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4" containerName="registry-server" Mar 21 05:38:00 crc kubenswrapper[4775]: I0321 05:38:00.147566 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4" containerName="registry-server" Mar 21 05:38:00 crc kubenswrapper[4775]: I0321 05:38:00.147761 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0b49c68-9fa7-444c-b3cd-4bef5eff4765" containerName="oc" Mar 21 05:38:00 crc kubenswrapper[4775]: I0321 05:38:00.147790 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7563d1d2-9c8c-4ae9-8748-cf7f6f2503c4" containerName="registry-server" Mar 21 05:38:00 crc kubenswrapper[4775]: I0321 05:38:00.148568 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567858-n842g" Mar 21 05:38:00 crc kubenswrapper[4775]: I0321 05:38:00.153775 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:38:00 crc kubenswrapper[4775]: I0321 05:38:00.153873 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:38:00 crc kubenswrapper[4775]: I0321 05:38:00.154445 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:38:00 crc kubenswrapper[4775]: I0321 05:38:00.159485 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567858-n842g"] Mar 21 05:38:00 crc kubenswrapper[4775]: I0321 05:38:00.323599 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2qkt\" (UniqueName: \"kubernetes.io/projected/b9b806cf-43d2-4304-8508-8f5af524cb37-kube-api-access-c2qkt\") pod \"auto-csr-approver-29567858-n842g\" (UID: \"b9b806cf-43d2-4304-8508-8f5af524cb37\") " pod="openshift-infra/auto-csr-approver-29567858-n842g" Mar 21 05:38:00 crc kubenswrapper[4775]: I0321 05:38:00.425677 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2qkt\" (UniqueName: \"kubernetes.io/projected/b9b806cf-43d2-4304-8508-8f5af524cb37-kube-api-access-c2qkt\") pod \"auto-csr-approver-29567858-n842g\" (UID: \"b9b806cf-43d2-4304-8508-8f5af524cb37\") " pod="openshift-infra/auto-csr-approver-29567858-n842g" Mar 21 05:38:00 crc kubenswrapper[4775]: I0321 05:38:00.445243 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2qkt\" (UniqueName: \"kubernetes.io/projected/b9b806cf-43d2-4304-8508-8f5af524cb37-kube-api-access-c2qkt\") pod \"auto-csr-approver-29567858-n842g\" (UID: \"b9b806cf-43d2-4304-8508-8f5af524cb37\") " pod="openshift-infra/auto-csr-approver-29567858-n842g" Mar 21 05:38:00 crc kubenswrapper[4775]: I0321 05:38:00.496775 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567858-n842g" Mar 21 05:38:00 crc kubenswrapper[4775]: I0321 05:38:00.925444 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567858-n842g"] Mar 21 05:38:00 crc kubenswrapper[4775]: I0321 05:38:00.934031 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 05:38:01 crc kubenswrapper[4775]: I0321 05:38:01.730553 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567858-n842g" event={"ID":"b9b806cf-43d2-4304-8508-8f5af524cb37","Type":"ContainerStarted","Data":"067d93dbd9bcbf0db5009f8884065250c82b7e0029e12b5809002041baf60978"} Mar 21 05:38:02 crc kubenswrapper[4775]: I0321 05:38:02.744133 4775 generic.go:334] "Generic (PLEG): container finished" podID="b9b806cf-43d2-4304-8508-8f5af524cb37" containerID="111e8b4e076f4628dda53e32e64a39739272ad46a623d86e91f2a428452f6fb0" exitCode=0 Mar 21 05:38:02 crc kubenswrapper[4775]: I0321 05:38:02.744259 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567858-n842g" event={"ID":"b9b806cf-43d2-4304-8508-8f5af524cb37","Type":"ContainerDied","Data":"111e8b4e076f4628dda53e32e64a39739272ad46a623d86e91f2a428452f6fb0"} Mar 21 05:38:04 crc kubenswrapper[4775]: I0321 05:38:04.212011 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567858-n842g" Mar 21 05:38:04 crc kubenswrapper[4775]: I0321 05:38:04.409561 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2qkt\" (UniqueName: \"kubernetes.io/projected/b9b806cf-43d2-4304-8508-8f5af524cb37-kube-api-access-c2qkt\") pod \"b9b806cf-43d2-4304-8508-8f5af524cb37\" (UID: \"b9b806cf-43d2-4304-8508-8f5af524cb37\") " Mar 21 05:38:04 crc kubenswrapper[4775]: I0321 05:38:04.417436 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9b806cf-43d2-4304-8508-8f5af524cb37-kube-api-access-c2qkt" (OuterVolumeSpecName: "kube-api-access-c2qkt") pod "b9b806cf-43d2-4304-8508-8f5af524cb37" (UID: "b9b806cf-43d2-4304-8508-8f5af524cb37"). InnerVolumeSpecName "kube-api-access-c2qkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:38:04 crc kubenswrapper[4775]: I0321 05:38:04.512647 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2qkt\" (UniqueName: \"kubernetes.io/projected/b9b806cf-43d2-4304-8508-8f5af524cb37-kube-api-access-c2qkt\") on node \"crc\" DevicePath \"\"" Mar 21 05:38:04 crc kubenswrapper[4775]: I0321 05:38:04.762515 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567858-n842g" event={"ID":"b9b806cf-43d2-4304-8508-8f5af524cb37","Type":"ContainerDied","Data":"067d93dbd9bcbf0db5009f8884065250c82b7e0029e12b5809002041baf60978"} Mar 21 05:38:04 crc kubenswrapper[4775]: I0321 05:38:04.762584 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="067d93dbd9bcbf0db5009f8884065250c82b7e0029e12b5809002041baf60978" Mar 21 05:38:04 crc kubenswrapper[4775]: I0321 05:38:04.762593 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567858-n842g" Mar 21 05:38:05 crc kubenswrapper[4775]: I0321 05:38:05.285106 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567852-hd7nd"] Mar 21 05:38:05 crc kubenswrapper[4775]: I0321 05:38:05.297522 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567852-hd7nd"] Mar 21 05:38:05 crc kubenswrapper[4775]: I0321 05:38:05.671566 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15017759-960b-4d6e-8ce9-deef2cb94155" path="/var/lib/kubelet/pods/15017759-960b-4d6e-8ce9-deef2cb94155/volumes" Mar 21 05:38:10 crc kubenswrapper[4775]: I0321 05:38:10.661767 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:38:10 crc kubenswrapper[4775]: E0321 05:38:10.662263 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:38:21 crc kubenswrapper[4775]: I0321 05:38:21.664257 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:38:21 crc kubenswrapper[4775]: E0321 05:38:21.666095 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:38:35 crc kubenswrapper[4775]: I0321 05:38:35.661232 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:38:35 crc kubenswrapper[4775]: E0321 05:38:35.662343 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:38:48 crc kubenswrapper[4775]: I0321 05:38:48.332038 4775 scope.go:117] "RemoveContainer" containerID="25828aaca134034962b25d7ea45e1870919d75c0c4923758d122b6c7bce2397c" Mar 21 05:38:49 crc kubenswrapper[4775]: I0321 05:38:49.661244 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:38:49 crc kubenswrapper[4775]: E0321 05:38:49.661580 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:39:03 crc kubenswrapper[4775]: I0321 05:39:03.662261 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:39:03 crc kubenswrapper[4775]: E0321 05:39:03.663111 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:39:16 crc kubenswrapper[4775]: I0321 05:39:16.662366 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:39:16 crc kubenswrapper[4775]: E0321 05:39:16.663880 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:39:29 crc kubenswrapper[4775]: I0321 05:39:29.661165 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:39:29 crc kubenswrapper[4775]: E0321 05:39:29.661900 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:39:42 crc kubenswrapper[4775]: I0321 05:39:42.662194 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:39:42 crc kubenswrapper[4775]: E0321 05:39:42.663146 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:39:57 crc kubenswrapper[4775]: I0321 05:39:57.667381 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:39:57 crc kubenswrapper[4775]: E0321 05:39:57.668260 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:40:00 crc kubenswrapper[4775]: I0321 05:40:00.156456 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567860-f4gsq"] Mar 21 05:40:00 crc kubenswrapper[4775]: E0321 05:40:00.157469 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9b806cf-43d2-4304-8508-8f5af524cb37" containerName="oc" Mar 21 05:40:00 crc kubenswrapper[4775]: I0321 05:40:00.157559 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9b806cf-43d2-4304-8508-8f5af524cb37" containerName="oc" Mar 21 05:40:00 crc kubenswrapper[4775]: I0321 05:40:00.158013 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9b806cf-43d2-4304-8508-8f5af524cb37" containerName="oc" Mar 21 05:40:00 crc kubenswrapper[4775]: I0321 05:40:00.158825 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567860-f4gsq" Mar 21 05:40:00 crc kubenswrapper[4775]: I0321 05:40:00.161219 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:40:00 crc kubenswrapper[4775]: I0321 05:40:00.161371 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:40:00 crc kubenswrapper[4775]: I0321 05:40:00.161392 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:40:00 crc kubenswrapper[4775]: I0321 05:40:00.167656 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567860-f4gsq"] Mar 21 05:40:00 crc kubenswrapper[4775]: I0321 05:40:00.312646 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhg6k\" (UniqueName: \"kubernetes.io/projected/42fbc24f-3528-4778-ae85-23e0c2fe99d9-kube-api-access-nhg6k\") pod \"auto-csr-approver-29567860-f4gsq\" (UID: \"42fbc24f-3528-4778-ae85-23e0c2fe99d9\") " pod="openshift-infra/auto-csr-approver-29567860-f4gsq" Mar 21 05:40:00 crc kubenswrapper[4775]: I0321 05:40:00.414511 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhg6k\" (UniqueName: \"kubernetes.io/projected/42fbc24f-3528-4778-ae85-23e0c2fe99d9-kube-api-access-nhg6k\") pod \"auto-csr-approver-29567860-f4gsq\" (UID: \"42fbc24f-3528-4778-ae85-23e0c2fe99d9\") " pod="openshift-infra/auto-csr-approver-29567860-f4gsq" Mar 21 05:40:00 crc kubenswrapper[4775]: I0321 05:40:00.450700 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhg6k\" (UniqueName: \"kubernetes.io/projected/42fbc24f-3528-4778-ae85-23e0c2fe99d9-kube-api-access-nhg6k\") pod \"auto-csr-approver-29567860-f4gsq\" (UID: \"42fbc24f-3528-4778-ae85-23e0c2fe99d9\") " pod="openshift-infra/auto-csr-approver-29567860-f4gsq" Mar 21 05:40:00 crc kubenswrapper[4775]: I0321 05:40:00.500312 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567860-f4gsq" Mar 21 05:40:01 crc kubenswrapper[4775]: I0321 05:40:01.000539 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567860-f4gsq"] Mar 21 05:40:01 crc kubenswrapper[4775]: I0321 05:40:01.026450 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567860-f4gsq" event={"ID":"42fbc24f-3528-4778-ae85-23e0c2fe99d9","Type":"ContainerStarted","Data":"6760cd7c21187c27e823e38f6472df2e0c87e6b8a85e2a72d82e6594ddfe9fc2"} Mar 21 05:40:04 crc kubenswrapper[4775]: I0321 05:40:04.052681 4775 generic.go:334] "Generic (PLEG): container finished" podID="42fbc24f-3528-4778-ae85-23e0c2fe99d9" containerID="772f50d7983f92c55a24cd77459f228985481da63abdb797f0dc2f295541ac61" exitCode=0 Mar 21 05:40:04 crc kubenswrapper[4775]: I0321 05:40:04.052776 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567860-f4gsq" event={"ID":"42fbc24f-3528-4778-ae85-23e0c2fe99d9","Type":"ContainerDied","Data":"772f50d7983f92c55a24cd77459f228985481da63abdb797f0dc2f295541ac61"} Mar 21 05:40:05 crc kubenswrapper[4775]: I0321 05:40:05.424684 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567860-f4gsq" Mar 21 05:40:05 crc kubenswrapper[4775]: I0321 05:40:05.449680 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhg6k\" (UniqueName: \"kubernetes.io/projected/42fbc24f-3528-4778-ae85-23e0c2fe99d9-kube-api-access-nhg6k\") pod \"42fbc24f-3528-4778-ae85-23e0c2fe99d9\" (UID: \"42fbc24f-3528-4778-ae85-23e0c2fe99d9\") " Mar 21 05:40:05 crc kubenswrapper[4775]: I0321 05:40:05.459666 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42fbc24f-3528-4778-ae85-23e0c2fe99d9-kube-api-access-nhg6k" (OuterVolumeSpecName: "kube-api-access-nhg6k") pod "42fbc24f-3528-4778-ae85-23e0c2fe99d9" (UID: "42fbc24f-3528-4778-ae85-23e0c2fe99d9"). InnerVolumeSpecName "kube-api-access-nhg6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:40:05 crc kubenswrapper[4775]: I0321 05:40:05.552355 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhg6k\" (UniqueName: \"kubernetes.io/projected/42fbc24f-3528-4778-ae85-23e0c2fe99d9-kube-api-access-nhg6k\") on node \"crc\" DevicePath \"\"" Mar 21 05:40:06 crc kubenswrapper[4775]: I0321 05:40:06.070969 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567860-f4gsq" event={"ID":"42fbc24f-3528-4778-ae85-23e0c2fe99d9","Type":"ContainerDied","Data":"6760cd7c21187c27e823e38f6472df2e0c87e6b8a85e2a72d82e6594ddfe9fc2"} Mar 21 05:40:06 crc kubenswrapper[4775]: I0321 05:40:06.071347 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6760cd7c21187c27e823e38f6472df2e0c87e6b8a85e2a72d82e6594ddfe9fc2" Mar 21 05:40:06 crc kubenswrapper[4775]: I0321 05:40:06.071038 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567860-f4gsq" Mar 21 05:40:06 crc kubenswrapper[4775]: I0321 05:40:06.530183 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567854-gxf82"] Mar 21 05:40:06 crc kubenswrapper[4775]: I0321 05:40:06.545047 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567854-gxf82"] Mar 21 05:40:07 crc kubenswrapper[4775]: I0321 05:40:07.673000 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afdb9c39-b2aa-496c-8cfb-917fd0b1cc15" path="/var/lib/kubelet/pods/afdb9c39-b2aa-496c-8cfb-917fd0b1cc15/volumes" Mar 21 05:40:10 crc kubenswrapper[4775]: I0321 05:40:10.661343 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:40:10 crc kubenswrapper[4775]: E0321 05:40:10.661890 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:40:25 crc kubenswrapper[4775]: I0321 05:40:25.661013 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:40:25 crc kubenswrapper[4775]: E0321 05:40:25.661880 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:40:37 crc kubenswrapper[4775]: I0321 05:40:37.688358 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:40:37 crc kubenswrapper[4775]: E0321 05:40:37.689415 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:40:47 crc kubenswrapper[4775]: I0321 05:40:47.676578 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-59krg"] Mar 21 05:40:47 crc kubenswrapper[4775]: E0321 05:40:47.677536 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42fbc24f-3528-4778-ae85-23e0c2fe99d9" containerName="oc" Mar 21 05:40:47 crc kubenswrapper[4775]: I0321 05:40:47.677558 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="42fbc24f-3528-4778-ae85-23e0c2fe99d9" containerName="oc" Mar 21 05:40:47 crc kubenswrapper[4775]: I0321 05:40:47.677866 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="42fbc24f-3528-4778-ae85-23e0c2fe99d9" containerName="oc" Mar 21 05:40:47 crc kubenswrapper[4775]: I0321 05:40:47.679848 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-59krg" Mar 21 05:40:47 crc kubenswrapper[4775]: I0321 05:40:47.686672 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-59krg"] Mar 21 05:40:47 crc kubenswrapper[4775]: I0321 05:40:47.745866 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c64d253c-50fa-492f-a3df-47eca811f134-catalog-content\") pod \"redhat-operators-59krg\" (UID: \"c64d253c-50fa-492f-a3df-47eca811f134\") " pod="openshift-marketplace/redhat-operators-59krg" Mar 21 05:40:47 crc kubenswrapper[4775]: I0321 05:40:47.746197 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c64d253c-50fa-492f-a3df-47eca811f134-utilities\") pod \"redhat-operators-59krg\" (UID: \"c64d253c-50fa-492f-a3df-47eca811f134\") " pod="openshift-marketplace/redhat-operators-59krg" Mar 21 05:40:47 crc kubenswrapper[4775]: I0321 05:40:47.746344 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghxxm\" (UniqueName: \"kubernetes.io/projected/c64d253c-50fa-492f-a3df-47eca811f134-kube-api-access-ghxxm\") pod \"redhat-operators-59krg\" (UID: \"c64d253c-50fa-492f-a3df-47eca811f134\") " pod="openshift-marketplace/redhat-operators-59krg" Mar 21 05:40:47 crc kubenswrapper[4775]: I0321 05:40:47.848450 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c64d253c-50fa-492f-a3df-47eca811f134-catalog-content\") pod \"redhat-operators-59krg\" (UID: \"c64d253c-50fa-492f-a3df-47eca811f134\") " pod="openshift-marketplace/redhat-operators-59krg" Mar 21 05:40:47 crc kubenswrapper[4775]: I0321 05:40:47.848601 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c64d253c-50fa-492f-a3df-47eca811f134-utilities\") pod \"redhat-operators-59krg\" (UID: \"c64d253c-50fa-492f-a3df-47eca811f134\") " pod="openshift-marketplace/redhat-operators-59krg" Mar 21 05:40:47 crc kubenswrapper[4775]: I0321 05:40:47.848649 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghxxm\" (UniqueName: \"kubernetes.io/projected/c64d253c-50fa-492f-a3df-47eca811f134-kube-api-access-ghxxm\") pod \"redhat-operators-59krg\" (UID: \"c64d253c-50fa-492f-a3df-47eca811f134\") " pod="openshift-marketplace/redhat-operators-59krg" Mar 21 05:40:47 crc kubenswrapper[4775]: I0321 05:40:47.848937 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c64d253c-50fa-492f-a3df-47eca811f134-catalog-content\") pod \"redhat-operators-59krg\" (UID: \"c64d253c-50fa-492f-a3df-47eca811f134\") " pod="openshift-marketplace/redhat-operators-59krg" Mar 21 05:40:47 crc kubenswrapper[4775]: I0321 05:40:47.849057 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c64d253c-50fa-492f-a3df-47eca811f134-utilities\") pod \"redhat-operators-59krg\" (UID: \"c64d253c-50fa-492f-a3df-47eca811f134\") " pod="openshift-marketplace/redhat-operators-59krg" Mar 21 05:40:47 crc kubenswrapper[4775]: I0321 05:40:47.904285 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghxxm\" (UniqueName: \"kubernetes.io/projected/c64d253c-50fa-492f-a3df-47eca811f134-kube-api-access-ghxxm\") pod \"redhat-operators-59krg\" (UID: \"c64d253c-50fa-492f-a3df-47eca811f134\") " pod="openshift-marketplace/redhat-operators-59krg" Mar 21 05:40:48 crc kubenswrapper[4775]: I0321 05:40:48.011153 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-59krg" Mar 21 05:40:48 crc kubenswrapper[4775]: I0321 05:40:48.421320 4775 scope.go:117] "RemoveContainer" containerID="20714a401828b9647e317e22423f1e191dc985cb13d154d0fa57a94896564896" Mar 21 05:40:48 crc kubenswrapper[4775]: I0321 05:40:48.496759 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-59krg"] Mar 21 05:40:49 crc kubenswrapper[4775]: I0321 05:40:49.479537 4775 generic.go:334] "Generic (PLEG): container finished" podID="c64d253c-50fa-492f-a3df-47eca811f134" containerID="700a025de9fada119b93b8005462e4c55304278415f6ec6840a7bd2dd4fe6a37" exitCode=0 Mar 21 05:40:49 crc kubenswrapper[4775]: I0321 05:40:49.479684 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59krg" event={"ID":"c64d253c-50fa-492f-a3df-47eca811f134","Type":"ContainerDied","Data":"700a025de9fada119b93b8005462e4c55304278415f6ec6840a7bd2dd4fe6a37"} Mar 21 05:40:49 crc kubenswrapper[4775]: I0321 05:40:49.480937 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59krg" event={"ID":"c64d253c-50fa-492f-a3df-47eca811f134","Type":"ContainerStarted","Data":"77c20c16df362ce9ee895d366c0649f371e0ba1c56bfd5984e139390c537c4cc"} Mar 21 05:40:52 crc kubenswrapper[4775]: I0321 05:40:52.511679 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59krg" event={"ID":"c64d253c-50fa-492f-a3df-47eca811f134","Type":"ContainerStarted","Data":"a3dfd9b083b6a3d2298d5af105729159a72ee6cc98844f82a4efca8eeecdcf38"} Mar 21 05:40:52 crc kubenswrapper[4775]: I0321 05:40:52.661649 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:40:52 crc kubenswrapper[4775]: E0321 05:40:52.662247 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:41:04 crc kubenswrapper[4775]: I0321 05:41:04.661824 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:41:04 crc kubenswrapper[4775]: E0321 05:41:04.662620 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:41:08 crc kubenswrapper[4775]: I0321 05:41:08.651458 4775 generic.go:334] "Generic (PLEG): container finished" podID="c64d253c-50fa-492f-a3df-47eca811f134" containerID="a3dfd9b083b6a3d2298d5af105729159a72ee6cc98844f82a4efca8eeecdcf38" exitCode=0 Mar 21 05:41:08 crc kubenswrapper[4775]: I0321 05:41:08.651535 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59krg" event={"ID":"c64d253c-50fa-492f-a3df-47eca811f134","Type":"ContainerDied","Data":"a3dfd9b083b6a3d2298d5af105729159a72ee6cc98844f82a4efca8eeecdcf38"} Mar 21 05:41:09 crc kubenswrapper[4775]: I0321 05:41:09.677641 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59krg" event={"ID":"c64d253c-50fa-492f-a3df-47eca811f134","Type":"ContainerStarted","Data":"aa6271b27072a7e8897fe55d4458fbaffe65abbbda75fce97d2afa7d3ec31cda"} Mar 21 05:41:09 crc kubenswrapper[4775]: I0321 05:41:09.707812 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-59krg" podStartSLOduration=3.079393263 podStartE2EDuration="22.707791283s" podCreationTimestamp="2026-03-21 05:40:47 +0000 UTC" firstStartedPulling="2026-03-21 05:40:49.482383295 +0000 UTC m=+3202.458846919" lastFinishedPulling="2026-03-21 05:41:09.110781315 +0000 UTC m=+3222.087244939" observedRunningTime="2026-03-21 05:41:09.700647599 +0000 UTC m=+3222.677111223" watchObservedRunningTime="2026-03-21 05:41:09.707791283 +0000 UTC m=+3222.684254907" Mar 21 05:41:18 crc kubenswrapper[4775]: I0321 05:41:18.011628 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-59krg" Mar 21 05:41:18 crc kubenswrapper[4775]: I0321 05:41:18.012564 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-59krg" Mar 21 05:41:18 crc kubenswrapper[4775]: I0321 05:41:18.066100 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-59krg" Mar 21 05:41:18 crc kubenswrapper[4775]: I0321 05:41:18.793057 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-59krg" Mar 21 05:41:18 crc kubenswrapper[4775]: I0321 05:41:18.860323 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-59krg"] Mar 21 05:41:19 crc kubenswrapper[4775]: I0321 05:41:19.662147 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:41:19 crc kubenswrapper[4775]: E0321 05:41:19.662608 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:41:20 crc kubenswrapper[4775]: I0321 05:41:20.761795 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-59krg" podUID="c64d253c-50fa-492f-a3df-47eca811f134" containerName="registry-server" containerID="cri-o://aa6271b27072a7e8897fe55d4458fbaffe65abbbda75fce97d2afa7d3ec31cda" gracePeriod=2 Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.431470 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-59krg" Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.479338 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghxxm\" (UniqueName: \"kubernetes.io/projected/c64d253c-50fa-492f-a3df-47eca811f134-kube-api-access-ghxxm\") pod \"c64d253c-50fa-492f-a3df-47eca811f134\" (UID: \"c64d253c-50fa-492f-a3df-47eca811f134\") " Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.479483 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c64d253c-50fa-492f-a3df-47eca811f134-catalog-content\") pod \"c64d253c-50fa-492f-a3df-47eca811f134\" (UID: \"c64d253c-50fa-492f-a3df-47eca811f134\") " Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.479575 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c64d253c-50fa-492f-a3df-47eca811f134-utilities\") pod \"c64d253c-50fa-492f-a3df-47eca811f134\" (UID: \"c64d253c-50fa-492f-a3df-47eca811f134\") " Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.480982 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c64d253c-50fa-492f-a3df-47eca811f134-utilities" (OuterVolumeSpecName: "utilities") pod "c64d253c-50fa-492f-a3df-47eca811f134" (UID: "c64d253c-50fa-492f-a3df-47eca811f134"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.486244 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c64d253c-50fa-492f-a3df-47eca811f134-kube-api-access-ghxxm" (OuterVolumeSpecName: "kube-api-access-ghxxm") pod "c64d253c-50fa-492f-a3df-47eca811f134" (UID: "c64d253c-50fa-492f-a3df-47eca811f134"). InnerVolumeSpecName "kube-api-access-ghxxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.581785 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c64d253c-50fa-492f-a3df-47eca811f134-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.581817 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghxxm\" (UniqueName: \"kubernetes.io/projected/c64d253c-50fa-492f-a3df-47eca811f134-kube-api-access-ghxxm\") on node \"crc\" DevicePath \"\"" Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.622422 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c64d253c-50fa-492f-a3df-47eca811f134-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c64d253c-50fa-492f-a3df-47eca811f134" (UID: "c64d253c-50fa-492f-a3df-47eca811f134"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.683957 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c64d253c-50fa-492f-a3df-47eca811f134-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.772147 4775 generic.go:334] "Generic (PLEG): container finished" podID="c64d253c-50fa-492f-a3df-47eca811f134" containerID="aa6271b27072a7e8897fe55d4458fbaffe65abbbda75fce97d2afa7d3ec31cda" exitCode=0 Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.772207 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59krg" event={"ID":"c64d253c-50fa-492f-a3df-47eca811f134","Type":"ContainerDied","Data":"aa6271b27072a7e8897fe55d4458fbaffe65abbbda75fce97d2afa7d3ec31cda"} Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.772284 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59krg" event={"ID":"c64d253c-50fa-492f-a3df-47eca811f134","Type":"ContainerDied","Data":"77c20c16df362ce9ee895d366c0649f371e0ba1c56bfd5984e139390c537c4cc"} Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.772227 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-59krg" Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.772307 4775 scope.go:117] "RemoveContainer" containerID="aa6271b27072a7e8897fe55d4458fbaffe65abbbda75fce97d2afa7d3ec31cda" Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.800142 4775 scope.go:117] "RemoveContainer" containerID="a3dfd9b083b6a3d2298d5af105729159a72ee6cc98844f82a4efca8eeecdcf38" Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.807946 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-59krg"] Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.817733 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-59krg"] Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.824380 4775 scope.go:117] "RemoveContainer" containerID="700a025de9fada119b93b8005462e4c55304278415f6ec6840a7bd2dd4fe6a37" Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.858997 4775 scope.go:117] "RemoveContainer" containerID="aa6271b27072a7e8897fe55d4458fbaffe65abbbda75fce97d2afa7d3ec31cda" Mar 21 05:41:21 crc kubenswrapper[4775]: E0321 05:41:21.859570 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa6271b27072a7e8897fe55d4458fbaffe65abbbda75fce97d2afa7d3ec31cda\": container with ID starting with aa6271b27072a7e8897fe55d4458fbaffe65abbbda75fce97d2afa7d3ec31cda not found: ID does not exist" containerID="aa6271b27072a7e8897fe55d4458fbaffe65abbbda75fce97d2afa7d3ec31cda" Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.859603 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa6271b27072a7e8897fe55d4458fbaffe65abbbda75fce97d2afa7d3ec31cda"} err="failed to get container status \"aa6271b27072a7e8897fe55d4458fbaffe65abbbda75fce97d2afa7d3ec31cda\": rpc error: code = NotFound desc = could not find container \"aa6271b27072a7e8897fe55d4458fbaffe65abbbda75fce97d2afa7d3ec31cda\": container with ID starting with aa6271b27072a7e8897fe55d4458fbaffe65abbbda75fce97d2afa7d3ec31cda not found: ID does not exist" Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.859642 4775 scope.go:117] "RemoveContainer" containerID="a3dfd9b083b6a3d2298d5af105729159a72ee6cc98844f82a4efca8eeecdcf38" Mar 21 05:41:21 crc kubenswrapper[4775]: E0321 05:41:21.859993 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3dfd9b083b6a3d2298d5af105729159a72ee6cc98844f82a4efca8eeecdcf38\": container with ID starting with a3dfd9b083b6a3d2298d5af105729159a72ee6cc98844f82a4efca8eeecdcf38 not found: ID does not exist" containerID="a3dfd9b083b6a3d2298d5af105729159a72ee6cc98844f82a4efca8eeecdcf38" Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.860023 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3dfd9b083b6a3d2298d5af105729159a72ee6cc98844f82a4efca8eeecdcf38"} err="failed to get container status \"a3dfd9b083b6a3d2298d5af105729159a72ee6cc98844f82a4efca8eeecdcf38\": rpc error: code = NotFound desc = could not find container \"a3dfd9b083b6a3d2298d5af105729159a72ee6cc98844f82a4efca8eeecdcf38\": container with ID starting with a3dfd9b083b6a3d2298d5af105729159a72ee6cc98844f82a4efca8eeecdcf38 not found: ID does not exist" Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.860044 4775 scope.go:117] "RemoveContainer" containerID="700a025de9fada119b93b8005462e4c55304278415f6ec6840a7bd2dd4fe6a37" Mar 21 05:41:21 crc kubenswrapper[4775]: E0321 05:41:21.860558 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"700a025de9fada119b93b8005462e4c55304278415f6ec6840a7bd2dd4fe6a37\": container with ID starting with 700a025de9fada119b93b8005462e4c55304278415f6ec6840a7bd2dd4fe6a37 not found: ID does not exist" containerID="700a025de9fada119b93b8005462e4c55304278415f6ec6840a7bd2dd4fe6a37" Mar 21 05:41:21 crc kubenswrapper[4775]: I0321 05:41:21.860698 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"700a025de9fada119b93b8005462e4c55304278415f6ec6840a7bd2dd4fe6a37"} err="failed to get container status \"700a025de9fada119b93b8005462e4c55304278415f6ec6840a7bd2dd4fe6a37\": rpc error: code = NotFound desc = could not find container \"700a025de9fada119b93b8005462e4c55304278415f6ec6840a7bd2dd4fe6a37\": container with ID starting with 700a025de9fada119b93b8005462e4c55304278415f6ec6840a7bd2dd4fe6a37 not found: ID does not exist" Mar 21 05:41:23 crc kubenswrapper[4775]: I0321 05:41:23.672791 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c64d253c-50fa-492f-a3df-47eca811f134" path="/var/lib/kubelet/pods/c64d253c-50fa-492f-a3df-47eca811f134/volumes" Mar 21 05:41:30 crc kubenswrapper[4775]: I0321 05:41:30.661985 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:41:30 crc kubenswrapper[4775]: E0321 05:41:30.662829 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:41:44 crc kubenswrapper[4775]: I0321 05:41:44.661480 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:41:44 crc kubenswrapper[4775]: E0321 05:41:44.662353 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:41:57 crc kubenswrapper[4775]: I0321 05:41:57.667472 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:41:57 crc kubenswrapper[4775]: E0321 05:41:57.668298 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:42:00 crc kubenswrapper[4775]: I0321 05:42:00.194864 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567862-bbnhp"] Mar 21 05:42:00 crc kubenswrapper[4775]: E0321 05:42:00.195862 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c64d253c-50fa-492f-a3df-47eca811f134" containerName="extract-content" Mar 21 05:42:00 crc kubenswrapper[4775]: I0321 05:42:00.195878 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c64d253c-50fa-492f-a3df-47eca811f134" containerName="extract-content" Mar 21 05:42:00 crc kubenswrapper[4775]: E0321 05:42:00.195926 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c64d253c-50fa-492f-a3df-47eca811f134" containerName="extract-utilities" Mar 21 05:42:00 crc kubenswrapper[4775]: I0321 05:42:00.195936 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c64d253c-50fa-492f-a3df-47eca811f134" containerName="extract-utilities" Mar 21 05:42:00 crc kubenswrapper[4775]: E0321 05:42:00.195953 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c64d253c-50fa-492f-a3df-47eca811f134" containerName="registry-server" Mar 21 05:42:00 crc kubenswrapper[4775]: I0321 05:42:00.195959 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c64d253c-50fa-492f-a3df-47eca811f134" containerName="registry-server" Mar 21 05:42:00 crc kubenswrapper[4775]: I0321 05:42:00.196176 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c64d253c-50fa-492f-a3df-47eca811f134" containerName="registry-server" Mar 21 05:42:00 crc kubenswrapper[4775]: I0321 05:42:00.196808 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567862-bbnhp" Mar 21 05:42:00 crc kubenswrapper[4775]: I0321 05:42:00.204456 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:42:00 crc kubenswrapper[4775]: I0321 05:42:00.205266 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:42:00 crc kubenswrapper[4775]: I0321 05:42:00.205355 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:42:00 crc kubenswrapper[4775]: I0321 05:42:00.210480 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567862-bbnhp"] Mar 21 05:42:00 crc kubenswrapper[4775]: I0321 05:42:00.317790 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5snjv\" (UniqueName: \"kubernetes.io/projected/89f5316a-d6df-4c37-b878-4f3ef82d837b-kube-api-access-5snjv\") pod \"auto-csr-approver-29567862-bbnhp\" (UID: \"89f5316a-d6df-4c37-b878-4f3ef82d837b\") " pod="openshift-infra/auto-csr-approver-29567862-bbnhp" Mar 21 05:42:00 crc kubenswrapper[4775]: I0321 05:42:00.420502 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5snjv\" (UniqueName: \"kubernetes.io/projected/89f5316a-d6df-4c37-b878-4f3ef82d837b-kube-api-access-5snjv\") pod \"auto-csr-approver-29567862-bbnhp\" (UID: \"89f5316a-d6df-4c37-b878-4f3ef82d837b\") " pod="openshift-infra/auto-csr-approver-29567862-bbnhp" Mar 21 05:42:00 crc kubenswrapper[4775]: I0321 05:42:00.448417 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5snjv\" (UniqueName: \"kubernetes.io/projected/89f5316a-d6df-4c37-b878-4f3ef82d837b-kube-api-access-5snjv\") pod \"auto-csr-approver-29567862-bbnhp\" (UID: \"89f5316a-d6df-4c37-b878-4f3ef82d837b\") " pod="openshift-infra/auto-csr-approver-29567862-bbnhp" Mar 21 05:42:00 crc kubenswrapper[4775]: I0321 05:42:00.880918 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567862-bbnhp" Mar 21 05:42:01 crc kubenswrapper[4775]: I0321 05:42:01.452380 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567862-bbnhp"] Mar 21 05:42:02 crc kubenswrapper[4775]: I0321 05:42:02.343621 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567862-bbnhp" event={"ID":"89f5316a-d6df-4c37-b878-4f3ef82d837b","Type":"ContainerStarted","Data":"112eb22a8eb1e402751c2cac58926bd161046e0f023e9a78cfdde548ce731b30"} Mar 21 05:42:03 crc kubenswrapper[4775]: I0321 05:42:03.354859 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567862-bbnhp" event={"ID":"89f5316a-d6df-4c37-b878-4f3ef82d837b","Type":"ContainerStarted","Data":"5c92c80d8920da595f41beeabde7d0200200f2dfb0bec0c216c4b59513f9ce40"} Mar 21 05:42:03 crc kubenswrapper[4775]: I0321 05:42:03.404842 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29567862-bbnhp" podStartSLOduration=1.861614872 podStartE2EDuration="3.404802468s" podCreationTimestamp="2026-03-21 05:42:00 +0000 UTC" firstStartedPulling="2026-03-21 05:42:01.460242244 +0000 UTC m=+3274.436705858" lastFinishedPulling="2026-03-21 05:42:03.00342983 +0000 UTC m=+3275.979893454" observedRunningTime="2026-03-21 05:42:03.393450564 +0000 UTC m=+3276.369914198" watchObservedRunningTime="2026-03-21 05:42:03.404802468 +0000 UTC m=+3276.381266092" Mar 21 05:42:04 crc kubenswrapper[4775]: I0321 05:42:04.363057 4775 generic.go:334] "Generic (PLEG): container finished" podID="89f5316a-d6df-4c37-b878-4f3ef82d837b" containerID="5c92c80d8920da595f41beeabde7d0200200f2dfb0bec0c216c4b59513f9ce40" exitCode=0 Mar 21 05:42:04 crc kubenswrapper[4775]: I0321 05:42:04.363222 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567862-bbnhp" event={"ID":"89f5316a-d6df-4c37-b878-4f3ef82d837b","Type":"ContainerDied","Data":"5c92c80d8920da595f41beeabde7d0200200f2dfb0bec0c216c4b59513f9ce40"} Mar 21 05:42:05 crc kubenswrapper[4775]: I0321 05:42:05.825909 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567862-bbnhp" Mar 21 05:42:05 crc kubenswrapper[4775]: I0321 05:42:05.935336 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5snjv\" (UniqueName: \"kubernetes.io/projected/89f5316a-d6df-4c37-b878-4f3ef82d837b-kube-api-access-5snjv\") pod \"89f5316a-d6df-4c37-b878-4f3ef82d837b\" (UID: \"89f5316a-d6df-4c37-b878-4f3ef82d837b\") " Mar 21 05:42:05 crc kubenswrapper[4775]: I0321 05:42:05.942691 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89f5316a-d6df-4c37-b878-4f3ef82d837b-kube-api-access-5snjv" (OuterVolumeSpecName: "kube-api-access-5snjv") pod "89f5316a-d6df-4c37-b878-4f3ef82d837b" (UID: "89f5316a-d6df-4c37-b878-4f3ef82d837b"). InnerVolumeSpecName "kube-api-access-5snjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:42:06 crc kubenswrapper[4775]: I0321 05:42:06.039342 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5snjv\" (UniqueName: \"kubernetes.io/projected/89f5316a-d6df-4c37-b878-4f3ef82d837b-kube-api-access-5snjv\") on node \"crc\" DevicePath \"\"" Mar 21 05:42:06 crc kubenswrapper[4775]: I0321 05:42:06.385465 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567862-bbnhp" event={"ID":"89f5316a-d6df-4c37-b878-4f3ef82d837b","Type":"ContainerDied","Data":"112eb22a8eb1e402751c2cac58926bd161046e0f023e9a78cfdde548ce731b30"} Mar 21 05:42:06 crc kubenswrapper[4775]: I0321 05:42:06.385953 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="112eb22a8eb1e402751c2cac58926bd161046e0f023e9a78cfdde548ce731b30" Mar 21 05:42:06 crc kubenswrapper[4775]: I0321 05:42:06.385705 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567862-bbnhp" Mar 21 05:42:06 crc kubenswrapper[4775]: I0321 05:42:06.898618 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567856-vjs55"] Mar 21 05:42:06 crc kubenswrapper[4775]: I0321 05:42:06.908315 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567856-vjs55"] Mar 21 05:42:07 crc kubenswrapper[4775]: I0321 05:42:07.672350 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0b49c68-9fa7-444c-b3cd-4bef5eff4765" path="/var/lib/kubelet/pods/d0b49c68-9fa7-444c-b3cd-4bef5eff4765/volumes" Mar 21 05:42:11 crc kubenswrapper[4775]: I0321 05:42:11.662724 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:42:11 crc kubenswrapper[4775]: E0321 05:42:11.663807 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:42:26 crc kubenswrapper[4775]: I0321 05:42:26.661958 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:42:26 crc kubenswrapper[4775]: E0321 05:42:26.662872 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:42:38 crc kubenswrapper[4775]: I0321 05:42:38.661524 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:42:39 crc kubenswrapper[4775]: I0321 05:42:39.716582 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerStarted","Data":"27384562242ad4f001715c32b1c4d680359f25816c963e4f026f07e0e28e85a2"} Mar 21 05:42:48 crc kubenswrapper[4775]: I0321 05:42:48.541333 4775 scope.go:117] "RemoveContainer" containerID="fe0d6a170f19e7078dcaeb489e7d9b4bba7aef3e240f5e59424f9dea5b0d382f" Mar 21 05:44:00 crc kubenswrapper[4775]: I0321 05:44:00.144810 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567864-zsztj"] Mar 21 05:44:00 crc kubenswrapper[4775]: E0321 05:44:00.145872 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89f5316a-d6df-4c37-b878-4f3ef82d837b" containerName="oc" Mar 21 05:44:00 crc kubenswrapper[4775]: I0321 05:44:00.145888 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="89f5316a-d6df-4c37-b878-4f3ef82d837b" containerName="oc" Mar 21 05:44:00 crc kubenswrapper[4775]: I0321 05:44:00.146080 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="89f5316a-d6df-4c37-b878-4f3ef82d837b" containerName="oc" Mar 21 05:44:00 crc kubenswrapper[4775]: I0321 05:44:00.146861 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567864-zsztj" Mar 21 05:44:00 crc kubenswrapper[4775]: I0321 05:44:00.149733 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:44:00 crc kubenswrapper[4775]: I0321 05:44:00.150411 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:44:00 crc kubenswrapper[4775]: I0321 05:44:00.150606 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:44:00 crc kubenswrapper[4775]: I0321 05:44:00.158409 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567864-zsztj"] Mar 21 05:44:00 crc kubenswrapper[4775]: I0321 05:44:00.316484 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhmtp\" (UniqueName: \"kubernetes.io/projected/3565ae50-5b89-4775-a8c2-45209b86619f-kube-api-access-zhmtp\") pod \"auto-csr-approver-29567864-zsztj\" (UID: \"3565ae50-5b89-4775-a8c2-45209b86619f\") " pod="openshift-infra/auto-csr-approver-29567864-zsztj" Mar 21 05:44:00 crc kubenswrapper[4775]: I0321 05:44:00.418339 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhmtp\" (UniqueName: \"kubernetes.io/projected/3565ae50-5b89-4775-a8c2-45209b86619f-kube-api-access-zhmtp\") pod \"auto-csr-approver-29567864-zsztj\" (UID: \"3565ae50-5b89-4775-a8c2-45209b86619f\") " pod="openshift-infra/auto-csr-approver-29567864-zsztj" Mar 21 05:44:00 crc kubenswrapper[4775]: I0321 05:44:00.441378 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhmtp\" (UniqueName: \"kubernetes.io/projected/3565ae50-5b89-4775-a8c2-45209b86619f-kube-api-access-zhmtp\") pod \"auto-csr-approver-29567864-zsztj\" (UID: \"3565ae50-5b89-4775-a8c2-45209b86619f\") " pod="openshift-infra/auto-csr-approver-29567864-zsztj" Mar 21 05:44:00 crc kubenswrapper[4775]: I0321 05:44:00.470350 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567864-zsztj" Mar 21 05:44:01 crc kubenswrapper[4775]: I0321 05:44:01.151599 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567864-zsztj"] Mar 21 05:44:01 crc kubenswrapper[4775]: W0321 05:44:01.161380 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3565ae50_5b89_4775_a8c2_45209b86619f.slice/crio-6a46803ca1a7c435f38c1681d6ceba3928009c92f6a9af994162f97a780be6c1 WatchSource:0}: Error finding container 6a46803ca1a7c435f38c1681d6ceba3928009c92f6a9af994162f97a780be6c1: Status 404 returned error can't find the container with id 6a46803ca1a7c435f38c1681d6ceba3928009c92f6a9af994162f97a780be6c1 Mar 21 05:44:01 crc kubenswrapper[4775]: I0321 05:44:01.163729 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 05:44:01 crc kubenswrapper[4775]: I0321 05:44:01.857501 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567864-zsztj" event={"ID":"3565ae50-5b89-4775-a8c2-45209b86619f","Type":"ContainerStarted","Data":"6a46803ca1a7c435f38c1681d6ceba3928009c92f6a9af994162f97a780be6c1"} Mar 21 05:44:02 crc kubenswrapper[4775]: I0321 05:44:02.870195 4775 generic.go:334] "Generic (PLEG): container finished" podID="3565ae50-5b89-4775-a8c2-45209b86619f" containerID="dbc52e5cee44a0750d2409abd45a8584ff150c5de4624d45b211f61e09e87efd" exitCode=0 Mar 21 05:44:02 crc kubenswrapper[4775]: I0321 05:44:02.870393 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567864-zsztj" event={"ID":"3565ae50-5b89-4775-a8c2-45209b86619f","Type":"ContainerDied","Data":"dbc52e5cee44a0750d2409abd45a8584ff150c5de4624d45b211f61e09e87efd"} Mar 21 05:44:04 crc kubenswrapper[4775]: I0321 05:44:04.395545 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567864-zsztj" Mar 21 05:44:04 crc kubenswrapper[4775]: I0321 05:44:04.406553 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhmtp\" (UniqueName: \"kubernetes.io/projected/3565ae50-5b89-4775-a8c2-45209b86619f-kube-api-access-zhmtp\") pod \"3565ae50-5b89-4775-a8c2-45209b86619f\" (UID: \"3565ae50-5b89-4775-a8c2-45209b86619f\") " Mar 21 05:44:04 crc kubenswrapper[4775]: I0321 05:44:04.420330 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3565ae50-5b89-4775-a8c2-45209b86619f-kube-api-access-zhmtp" (OuterVolumeSpecName: "kube-api-access-zhmtp") pod "3565ae50-5b89-4775-a8c2-45209b86619f" (UID: "3565ae50-5b89-4775-a8c2-45209b86619f"). InnerVolumeSpecName "kube-api-access-zhmtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:44:04 crc kubenswrapper[4775]: I0321 05:44:04.509048 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhmtp\" (UniqueName: \"kubernetes.io/projected/3565ae50-5b89-4775-a8c2-45209b86619f-kube-api-access-zhmtp\") on node \"crc\" DevicePath \"\"" Mar 21 05:44:04 crc kubenswrapper[4775]: I0321 05:44:04.888492 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567864-zsztj" event={"ID":"3565ae50-5b89-4775-a8c2-45209b86619f","Type":"ContainerDied","Data":"6a46803ca1a7c435f38c1681d6ceba3928009c92f6a9af994162f97a780be6c1"} Mar 21 05:44:04 crc kubenswrapper[4775]: I0321 05:44:04.888540 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a46803ca1a7c435f38c1681d6ceba3928009c92f6a9af994162f97a780be6c1" Mar 21 05:44:04 crc kubenswrapper[4775]: I0321 05:44:04.888578 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567864-zsztj" Mar 21 05:44:05 crc kubenswrapper[4775]: I0321 05:44:05.478549 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567858-n842g"] Mar 21 05:44:05 crc kubenswrapper[4775]: I0321 05:44:05.490696 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567858-n842g"] Mar 21 05:44:05 crc kubenswrapper[4775]: I0321 05:44:05.671422 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9b806cf-43d2-4304-8508-8f5af524cb37" path="/var/lib/kubelet/pods/b9b806cf-43d2-4304-8508-8f5af524cb37/volumes" Mar 21 05:44:19 crc kubenswrapper[4775]: I0321 05:44:19.760807 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xpg9p"] Mar 21 05:44:19 crc kubenswrapper[4775]: E0321 05:44:19.761662 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3565ae50-5b89-4775-a8c2-45209b86619f" containerName="oc" Mar 21 05:44:19 crc kubenswrapper[4775]: I0321 05:44:19.761680 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3565ae50-5b89-4775-a8c2-45209b86619f" containerName="oc" Mar 21 05:44:19 crc kubenswrapper[4775]: I0321 05:44:19.761941 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3565ae50-5b89-4775-a8c2-45209b86619f" containerName="oc" Mar 21 05:44:19 crc kubenswrapper[4775]: I0321 05:44:19.763771 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xpg9p" Mar 21 05:44:19 crc kubenswrapper[4775]: I0321 05:44:19.791315 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xpg9p"] Mar 21 05:44:19 crc kubenswrapper[4775]: I0321 05:44:19.909267 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjmfs\" (UniqueName: \"kubernetes.io/projected/c328ad3c-41c4-419e-88ea-8bcd758dcb9e-kube-api-access-mjmfs\") pod \"certified-operators-xpg9p\" (UID: \"c328ad3c-41c4-419e-88ea-8bcd758dcb9e\") " pod="openshift-marketplace/certified-operators-xpg9p" Mar 21 05:44:19 crc kubenswrapper[4775]: I0321 05:44:19.909744 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c328ad3c-41c4-419e-88ea-8bcd758dcb9e-catalog-content\") pod \"certified-operators-xpg9p\" (UID: \"c328ad3c-41c4-419e-88ea-8bcd758dcb9e\") " pod="openshift-marketplace/certified-operators-xpg9p" Mar 21 05:44:19 crc kubenswrapper[4775]: I0321 05:44:19.909794 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c328ad3c-41c4-419e-88ea-8bcd758dcb9e-utilities\") pod \"certified-operators-xpg9p\" (UID: \"c328ad3c-41c4-419e-88ea-8bcd758dcb9e\") " pod="openshift-marketplace/certified-operators-xpg9p" Mar 21 05:44:20 crc kubenswrapper[4775]: I0321 05:44:20.011656 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c328ad3c-41c4-419e-88ea-8bcd758dcb9e-catalog-content\") pod \"certified-operators-xpg9p\" (UID: \"c328ad3c-41c4-419e-88ea-8bcd758dcb9e\") " pod="openshift-marketplace/certified-operators-xpg9p" Mar 21 05:44:20 crc kubenswrapper[4775]: I0321 05:44:20.011962 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c328ad3c-41c4-419e-88ea-8bcd758dcb9e-utilities\") pod \"certified-operators-xpg9p\" (UID: \"c328ad3c-41c4-419e-88ea-8bcd758dcb9e\") " pod="openshift-marketplace/certified-operators-xpg9p" Mar 21 05:44:20 crc kubenswrapper[4775]: I0321 05:44:20.012131 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjmfs\" (UniqueName: \"kubernetes.io/projected/c328ad3c-41c4-419e-88ea-8bcd758dcb9e-kube-api-access-mjmfs\") pod \"certified-operators-xpg9p\" (UID: \"c328ad3c-41c4-419e-88ea-8bcd758dcb9e\") " pod="openshift-marketplace/certified-operators-xpg9p" Mar 21 05:44:20 crc kubenswrapper[4775]: I0321 05:44:20.012340 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c328ad3c-41c4-419e-88ea-8bcd758dcb9e-catalog-content\") pod \"certified-operators-xpg9p\" (UID: \"c328ad3c-41c4-419e-88ea-8bcd758dcb9e\") " pod="openshift-marketplace/certified-operators-xpg9p" Mar 21 05:44:20 crc kubenswrapper[4775]: I0321 05:44:20.012440 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c328ad3c-41c4-419e-88ea-8bcd758dcb9e-utilities\") pod \"certified-operators-xpg9p\" (UID: \"c328ad3c-41c4-419e-88ea-8bcd758dcb9e\") " pod="openshift-marketplace/certified-operators-xpg9p" Mar 21 05:44:20 crc kubenswrapper[4775]: I0321 05:44:20.038205 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjmfs\" (UniqueName: \"kubernetes.io/projected/c328ad3c-41c4-419e-88ea-8bcd758dcb9e-kube-api-access-mjmfs\") pod \"certified-operators-xpg9p\" (UID: \"c328ad3c-41c4-419e-88ea-8bcd758dcb9e\") " pod="openshift-marketplace/certified-operators-xpg9p" Mar 21 05:44:20 crc kubenswrapper[4775]: I0321 05:44:20.084900 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xpg9p" Mar 21 05:44:20 crc kubenswrapper[4775]: I0321 05:44:20.706864 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xpg9p"] Mar 21 05:44:21 crc kubenswrapper[4775]: I0321 05:44:21.230765 4775 generic.go:334] "Generic (PLEG): container finished" podID="c328ad3c-41c4-419e-88ea-8bcd758dcb9e" containerID="9fb0be2f280c6c43746b5afba600c3d334f8a86d80cb113c8262eb021a4d9dad" exitCode=0 Mar 21 05:44:21 crc kubenswrapper[4775]: I0321 05:44:21.230825 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xpg9p" event={"ID":"c328ad3c-41c4-419e-88ea-8bcd758dcb9e","Type":"ContainerDied","Data":"9fb0be2f280c6c43746b5afba600c3d334f8a86d80cb113c8262eb021a4d9dad"} Mar 21 05:44:21 crc kubenswrapper[4775]: I0321 05:44:21.231053 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xpg9p" event={"ID":"c328ad3c-41c4-419e-88ea-8bcd758dcb9e","Type":"ContainerStarted","Data":"1555376c25e05b9026c4e9c7682126d2efcfb8571cf0c2c03216cdf17dee8a8d"} Mar 21 05:44:23 crc kubenswrapper[4775]: I0321 05:44:23.254591 4775 generic.go:334] "Generic (PLEG): container finished" podID="c328ad3c-41c4-419e-88ea-8bcd758dcb9e" containerID="e5c7b15bf2740b46601cb7a23aa579db3b92a3ad5e9f07f7746cfac8bf93d1bf" exitCode=0 Mar 21 05:44:23 crc kubenswrapper[4775]: I0321 05:44:23.254807 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xpg9p" event={"ID":"c328ad3c-41c4-419e-88ea-8bcd758dcb9e","Type":"ContainerDied","Data":"e5c7b15bf2740b46601cb7a23aa579db3b92a3ad5e9f07f7746cfac8bf93d1bf"} Mar 21 05:44:25 crc kubenswrapper[4775]: I0321 05:44:25.275977 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xpg9p" event={"ID":"c328ad3c-41c4-419e-88ea-8bcd758dcb9e","Type":"ContainerStarted","Data":"a5d4a873d066150b522769e8838cafe2ee9965846f80ed997ebe148b5055e6bf"} Mar 21 05:44:25 crc kubenswrapper[4775]: I0321 05:44:25.291701 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xpg9p" podStartSLOduration=3.179261875 podStartE2EDuration="6.291685837s" podCreationTimestamp="2026-03-21 05:44:19 +0000 UTC" firstStartedPulling="2026-03-21 05:44:21.232322256 +0000 UTC m=+3414.208785880" lastFinishedPulling="2026-03-21 05:44:24.344746218 +0000 UTC m=+3417.321209842" observedRunningTime="2026-03-21 05:44:25.290964226 +0000 UTC m=+3418.267427850" watchObservedRunningTime="2026-03-21 05:44:25.291685837 +0000 UTC m=+3418.268149461" Mar 21 05:44:30 crc kubenswrapper[4775]: I0321 05:44:30.086830 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xpg9p" Mar 21 05:44:30 crc kubenswrapper[4775]: I0321 05:44:30.087553 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xpg9p" Mar 21 05:44:30 crc kubenswrapper[4775]: I0321 05:44:30.137839 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xpg9p" Mar 21 05:44:30 crc kubenswrapper[4775]: I0321 05:44:30.368276 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xpg9p" Mar 21 05:44:30 crc kubenswrapper[4775]: I0321 05:44:30.425272 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xpg9p"] Mar 21 05:44:32 crc kubenswrapper[4775]: I0321 05:44:32.334850 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xpg9p" podUID="c328ad3c-41c4-419e-88ea-8bcd758dcb9e" containerName="registry-server" containerID="cri-o://a5d4a873d066150b522769e8838cafe2ee9965846f80ed997ebe148b5055e6bf" gracePeriod=2 Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.283676 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xpg9p" Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.424842 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c328ad3c-41c4-419e-88ea-8bcd758dcb9e-utilities\") pod \"c328ad3c-41c4-419e-88ea-8bcd758dcb9e\" (UID: \"c328ad3c-41c4-419e-88ea-8bcd758dcb9e\") " Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.425000 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c328ad3c-41c4-419e-88ea-8bcd758dcb9e-catalog-content\") pod \"c328ad3c-41c4-419e-88ea-8bcd758dcb9e\" (UID: \"c328ad3c-41c4-419e-88ea-8bcd758dcb9e\") " Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.425065 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjmfs\" (UniqueName: \"kubernetes.io/projected/c328ad3c-41c4-419e-88ea-8bcd758dcb9e-kube-api-access-mjmfs\") pod \"c328ad3c-41c4-419e-88ea-8bcd758dcb9e\" (UID: \"c328ad3c-41c4-419e-88ea-8bcd758dcb9e\") " Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.426340 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c328ad3c-41c4-419e-88ea-8bcd758dcb9e-utilities" (OuterVolumeSpecName: "utilities") pod "c328ad3c-41c4-419e-88ea-8bcd758dcb9e" (UID: "c328ad3c-41c4-419e-88ea-8bcd758dcb9e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.429206 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c328ad3c-41c4-419e-88ea-8bcd758dcb9e-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.430402 4775 generic.go:334] "Generic (PLEG): container finished" podID="c328ad3c-41c4-419e-88ea-8bcd758dcb9e" containerID="a5d4a873d066150b522769e8838cafe2ee9965846f80ed997ebe148b5055e6bf" exitCode=0 Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.430434 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xpg9p" event={"ID":"c328ad3c-41c4-419e-88ea-8bcd758dcb9e","Type":"ContainerDied","Data":"a5d4a873d066150b522769e8838cafe2ee9965846f80ed997ebe148b5055e6bf"} Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.430533 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xpg9p" event={"ID":"c328ad3c-41c4-419e-88ea-8bcd758dcb9e","Type":"ContainerDied","Data":"1555376c25e05b9026c4e9c7682126d2efcfb8571cf0c2c03216cdf17dee8a8d"} Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.430555 4775 scope.go:117] "RemoveContainer" containerID="a5d4a873d066150b522769e8838cafe2ee9965846f80ed997ebe148b5055e6bf" Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.430756 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xpg9p" Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.447774 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c328ad3c-41c4-419e-88ea-8bcd758dcb9e-kube-api-access-mjmfs" (OuterVolumeSpecName: "kube-api-access-mjmfs") pod "c328ad3c-41c4-419e-88ea-8bcd758dcb9e" (UID: "c328ad3c-41c4-419e-88ea-8bcd758dcb9e"). InnerVolumeSpecName "kube-api-access-mjmfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.502629 4775 scope.go:117] "RemoveContainer" containerID="e5c7b15bf2740b46601cb7a23aa579db3b92a3ad5e9f07f7746cfac8bf93d1bf" Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.511974 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c328ad3c-41c4-419e-88ea-8bcd758dcb9e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c328ad3c-41c4-419e-88ea-8bcd758dcb9e" (UID: "c328ad3c-41c4-419e-88ea-8bcd758dcb9e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.521355 4775 scope.go:117] "RemoveContainer" containerID="9fb0be2f280c6c43746b5afba600c3d334f8a86d80cb113c8262eb021a4d9dad" Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.530923 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c328ad3c-41c4-419e-88ea-8bcd758dcb9e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.531157 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjmfs\" (UniqueName: \"kubernetes.io/projected/c328ad3c-41c4-419e-88ea-8bcd758dcb9e-kube-api-access-mjmfs\") on node \"crc\" DevicePath \"\"" Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.598358 4775 scope.go:117] "RemoveContainer" containerID="a5d4a873d066150b522769e8838cafe2ee9965846f80ed997ebe148b5055e6bf" Mar 21 05:44:33 crc kubenswrapper[4775]: E0321 05:44:33.600590 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5d4a873d066150b522769e8838cafe2ee9965846f80ed997ebe148b5055e6bf\": container with ID starting with a5d4a873d066150b522769e8838cafe2ee9965846f80ed997ebe148b5055e6bf not found: ID does not exist" containerID="a5d4a873d066150b522769e8838cafe2ee9965846f80ed997ebe148b5055e6bf" Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.600701 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5d4a873d066150b522769e8838cafe2ee9965846f80ed997ebe148b5055e6bf"} err="failed to get container status \"a5d4a873d066150b522769e8838cafe2ee9965846f80ed997ebe148b5055e6bf\": rpc error: code = NotFound desc = could not find container \"a5d4a873d066150b522769e8838cafe2ee9965846f80ed997ebe148b5055e6bf\": container with ID starting with a5d4a873d066150b522769e8838cafe2ee9965846f80ed997ebe148b5055e6bf not found: ID does not exist" Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.600801 4775 scope.go:117] "RemoveContainer" containerID="e5c7b15bf2740b46601cb7a23aa579db3b92a3ad5e9f07f7746cfac8bf93d1bf" Mar 21 05:44:33 crc kubenswrapper[4775]: E0321 05:44:33.601077 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5c7b15bf2740b46601cb7a23aa579db3b92a3ad5e9f07f7746cfac8bf93d1bf\": container with ID starting with e5c7b15bf2740b46601cb7a23aa579db3b92a3ad5e9f07f7746cfac8bf93d1bf not found: ID does not exist" containerID="e5c7b15bf2740b46601cb7a23aa579db3b92a3ad5e9f07f7746cfac8bf93d1bf" Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.601188 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5c7b15bf2740b46601cb7a23aa579db3b92a3ad5e9f07f7746cfac8bf93d1bf"} err="failed to get container status \"e5c7b15bf2740b46601cb7a23aa579db3b92a3ad5e9f07f7746cfac8bf93d1bf\": rpc error: code = NotFound desc = could not find container \"e5c7b15bf2740b46601cb7a23aa579db3b92a3ad5e9f07f7746cfac8bf93d1bf\": container with ID starting with e5c7b15bf2740b46601cb7a23aa579db3b92a3ad5e9f07f7746cfac8bf93d1bf not found: ID does not exist" Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.601281 4775 scope.go:117] "RemoveContainer" containerID="9fb0be2f280c6c43746b5afba600c3d334f8a86d80cb113c8262eb021a4d9dad" Mar 21 05:44:33 crc kubenswrapper[4775]: E0321 05:44:33.603346 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fb0be2f280c6c43746b5afba600c3d334f8a86d80cb113c8262eb021a4d9dad\": container with ID starting with 9fb0be2f280c6c43746b5afba600c3d334f8a86d80cb113c8262eb021a4d9dad not found: ID does not exist" containerID="9fb0be2f280c6c43746b5afba600c3d334f8a86d80cb113c8262eb021a4d9dad" Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.603404 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fb0be2f280c6c43746b5afba600c3d334f8a86d80cb113c8262eb021a4d9dad"} err="failed to get container status \"9fb0be2f280c6c43746b5afba600c3d334f8a86d80cb113c8262eb021a4d9dad\": rpc error: code = NotFound desc = could not find container \"9fb0be2f280c6c43746b5afba600c3d334f8a86d80cb113c8262eb021a4d9dad\": container with ID starting with 9fb0be2f280c6c43746b5afba600c3d334f8a86d80cb113c8262eb021a4d9dad not found: ID does not exist" Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.757341 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xpg9p"] Mar 21 05:44:33 crc kubenswrapper[4775]: I0321 05:44:33.769574 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xpg9p"] Mar 21 05:44:35 crc kubenswrapper[4775]: I0321 05:44:35.670847 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c328ad3c-41c4-419e-88ea-8bcd758dcb9e" path="/var/lib/kubelet/pods/c328ad3c-41c4-419e-88ea-8bcd758dcb9e/volumes" Mar 21 05:44:48 crc kubenswrapper[4775]: I0321 05:44:48.655790 4775 scope.go:117] "RemoveContainer" containerID="111e8b4e076f4628dda53e32e64a39739272ad46a623d86e91f2a428452f6fb0" Mar 21 05:45:00 crc kubenswrapper[4775]: I0321 05:45:00.150113 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm"] Mar 21 05:45:00 crc kubenswrapper[4775]: E0321 05:45:00.151093 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c328ad3c-41c4-419e-88ea-8bcd758dcb9e" containerName="extract-content" Mar 21 05:45:00 crc kubenswrapper[4775]: I0321 05:45:00.151107 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c328ad3c-41c4-419e-88ea-8bcd758dcb9e" containerName="extract-content" Mar 21 05:45:00 crc kubenswrapper[4775]: E0321 05:45:00.151153 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c328ad3c-41c4-419e-88ea-8bcd758dcb9e" containerName="extract-utilities" Mar 21 05:45:00 crc kubenswrapper[4775]: I0321 05:45:00.151160 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c328ad3c-41c4-419e-88ea-8bcd758dcb9e" containerName="extract-utilities" Mar 21 05:45:00 crc kubenswrapper[4775]: E0321 05:45:00.151178 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c328ad3c-41c4-419e-88ea-8bcd758dcb9e" containerName="registry-server" Mar 21 05:45:00 crc kubenswrapper[4775]: I0321 05:45:00.151184 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c328ad3c-41c4-419e-88ea-8bcd758dcb9e" containerName="registry-server" Mar 21 05:45:00 crc kubenswrapper[4775]: I0321 05:45:00.151354 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c328ad3c-41c4-419e-88ea-8bcd758dcb9e" containerName="registry-server" Mar 21 05:45:00 crc kubenswrapper[4775]: I0321 05:45:00.152035 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm" Mar 21 05:45:00 crc kubenswrapper[4775]: I0321 05:45:00.155147 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 21 05:45:00 crc kubenswrapper[4775]: I0321 05:45:00.155793 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 21 05:45:00 crc kubenswrapper[4775]: I0321 05:45:00.164768 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm"] Mar 21 05:45:00 crc kubenswrapper[4775]: I0321 05:45:00.274091 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnmvv\" (UniqueName: \"kubernetes.io/projected/853b684f-d127-48b8-9dd6-5eed4edccfe6-kube-api-access-hnmvv\") pod \"collect-profiles-29567865-dtkzm\" (UID: \"853b684f-d127-48b8-9dd6-5eed4edccfe6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm" Mar 21 05:45:00 crc kubenswrapper[4775]: I0321 05:45:00.274310 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/853b684f-d127-48b8-9dd6-5eed4edccfe6-secret-volume\") pod \"collect-profiles-29567865-dtkzm\" (UID: \"853b684f-d127-48b8-9dd6-5eed4edccfe6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm" Mar 21 05:45:00 crc kubenswrapper[4775]: I0321 05:45:00.274390 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/853b684f-d127-48b8-9dd6-5eed4edccfe6-config-volume\") pod \"collect-profiles-29567865-dtkzm\" (UID: \"853b684f-d127-48b8-9dd6-5eed4edccfe6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm" Mar 21 05:45:00 crc kubenswrapper[4775]: I0321 05:45:00.376526 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/853b684f-d127-48b8-9dd6-5eed4edccfe6-secret-volume\") pod \"collect-profiles-29567865-dtkzm\" (UID: \"853b684f-d127-48b8-9dd6-5eed4edccfe6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm" Mar 21 05:45:00 crc kubenswrapper[4775]: I0321 05:45:00.376672 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/853b684f-d127-48b8-9dd6-5eed4edccfe6-config-volume\") pod \"collect-profiles-29567865-dtkzm\" (UID: \"853b684f-d127-48b8-9dd6-5eed4edccfe6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm" Mar 21 05:45:00 crc kubenswrapper[4775]: I0321 05:45:00.376826 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnmvv\" (UniqueName: \"kubernetes.io/projected/853b684f-d127-48b8-9dd6-5eed4edccfe6-kube-api-access-hnmvv\") pod \"collect-profiles-29567865-dtkzm\" (UID: \"853b684f-d127-48b8-9dd6-5eed4edccfe6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm" Mar 21 05:45:00 crc kubenswrapper[4775]: I0321 05:45:00.378161 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/853b684f-d127-48b8-9dd6-5eed4edccfe6-config-volume\") pod \"collect-profiles-29567865-dtkzm\" (UID: \"853b684f-d127-48b8-9dd6-5eed4edccfe6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm" Mar 21 05:45:00 crc kubenswrapper[4775]: I0321 05:45:00.384214 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/853b684f-d127-48b8-9dd6-5eed4edccfe6-secret-volume\") pod \"collect-profiles-29567865-dtkzm\" (UID: \"853b684f-d127-48b8-9dd6-5eed4edccfe6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm" Mar 21 05:45:00 crc kubenswrapper[4775]: I0321 05:45:00.394609 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnmvv\" (UniqueName: \"kubernetes.io/projected/853b684f-d127-48b8-9dd6-5eed4edccfe6-kube-api-access-hnmvv\") pod \"collect-profiles-29567865-dtkzm\" (UID: \"853b684f-d127-48b8-9dd6-5eed4edccfe6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm" Mar 21 05:45:00 crc kubenswrapper[4775]: I0321 05:45:00.484871 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm" Mar 21 05:45:00 crc kubenswrapper[4775]: I0321 05:45:00.964317 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm"] Mar 21 05:45:01 crc kubenswrapper[4775]: I0321 05:45:01.145180 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm" event={"ID":"853b684f-d127-48b8-9dd6-5eed4edccfe6","Type":"ContainerStarted","Data":"1b4c8b22e4aa0409d0b120e79806d2e1fd24b5cce756221596dd023358442451"} Mar 21 05:45:01 crc kubenswrapper[4775]: I0321 05:45:01.145249 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm" event={"ID":"853b684f-d127-48b8-9dd6-5eed4edccfe6","Type":"ContainerStarted","Data":"a47c618d2b09bdc0b59e70299cd03eb578c080155feeb1dd527de0cec62c45e6"} Mar 21 05:45:01 crc kubenswrapper[4775]: I0321 05:45:01.167795 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm" podStartSLOduration=1.167775832 podStartE2EDuration="1.167775832s" podCreationTimestamp="2026-03-21 05:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:45:01.161628117 +0000 UTC m=+3454.138091741" watchObservedRunningTime="2026-03-21 05:45:01.167775832 +0000 UTC m=+3454.144239456" Mar 21 05:45:02 crc kubenswrapper[4775]: I0321 05:45:02.157853 4775 generic.go:334] "Generic (PLEG): container finished" podID="853b684f-d127-48b8-9dd6-5eed4edccfe6" containerID="1b4c8b22e4aa0409d0b120e79806d2e1fd24b5cce756221596dd023358442451" exitCode=0 Mar 21 05:45:02 crc kubenswrapper[4775]: I0321 05:45:02.157940 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm" event={"ID":"853b684f-d127-48b8-9dd6-5eed4edccfe6","Type":"ContainerDied","Data":"1b4c8b22e4aa0409d0b120e79806d2e1fd24b5cce756221596dd023358442451"} Mar 21 05:45:02 crc kubenswrapper[4775]: I0321 05:45:02.482974 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:45:02 crc kubenswrapper[4775]: I0321 05:45:02.483064 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:45:03 crc kubenswrapper[4775]: I0321 05:45:03.563616 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm" Mar 21 05:45:03 crc kubenswrapper[4775]: I0321 05:45:03.648279 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnmvv\" (UniqueName: \"kubernetes.io/projected/853b684f-d127-48b8-9dd6-5eed4edccfe6-kube-api-access-hnmvv\") pod \"853b684f-d127-48b8-9dd6-5eed4edccfe6\" (UID: \"853b684f-d127-48b8-9dd6-5eed4edccfe6\") " Mar 21 05:45:03 crc kubenswrapper[4775]: I0321 05:45:03.648339 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/853b684f-d127-48b8-9dd6-5eed4edccfe6-config-volume\") pod \"853b684f-d127-48b8-9dd6-5eed4edccfe6\" (UID: \"853b684f-d127-48b8-9dd6-5eed4edccfe6\") " Mar 21 05:45:03 crc kubenswrapper[4775]: I0321 05:45:03.648426 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/853b684f-d127-48b8-9dd6-5eed4edccfe6-secret-volume\") pod \"853b684f-d127-48b8-9dd6-5eed4edccfe6\" (UID: \"853b684f-d127-48b8-9dd6-5eed4edccfe6\") " Mar 21 05:45:03 crc kubenswrapper[4775]: I0321 05:45:03.649968 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/853b684f-d127-48b8-9dd6-5eed4edccfe6-config-volume" (OuterVolumeSpecName: "config-volume") pod "853b684f-d127-48b8-9dd6-5eed4edccfe6" (UID: "853b684f-d127-48b8-9dd6-5eed4edccfe6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:45:03 crc kubenswrapper[4775]: I0321 05:45:03.656652 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/853b684f-d127-48b8-9dd6-5eed4edccfe6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "853b684f-d127-48b8-9dd6-5eed4edccfe6" (UID: "853b684f-d127-48b8-9dd6-5eed4edccfe6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:45:03 crc kubenswrapper[4775]: I0321 05:45:03.656695 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/853b684f-d127-48b8-9dd6-5eed4edccfe6-kube-api-access-hnmvv" (OuterVolumeSpecName: "kube-api-access-hnmvv") pod "853b684f-d127-48b8-9dd6-5eed4edccfe6" (UID: "853b684f-d127-48b8-9dd6-5eed4edccfe6"). InnerVolumeSpecName "kube-api-access-hnmvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:45:03 crc kubenswrapper[4775]: I0321 05:45:03.750679 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/853b684f-d127-48b8-9dd6-5eed4edccfe6-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 21 05:45:03 crc kubenswrapper[4775]: I0321 05:45:03.751045 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnmvv\" (UniqueName: \"kubernetes.io/projected/853b684f-d127-48b8-9dd6-5eed4edccfe6-kube-api-access-hnmvv\") on node \"crc\" DevicePath \"\"" Mar 21 05:45:03 crc kubenswrapper[4775]: I0321 05:45:03.751316 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/853b684f-d127-48b8-9dd6-5eed4edccfe6-config-volume\") on node \"crc\" DevicePath \"\"" Mar 21 05:45:04 crc kubenswrapper[4775]: I0321 05:45:04.181795 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm" event={"ID":"853b684f-d127-48b8-9dd6-5eed4edccfe6","Type":"ContainerDied","Data":"a47c618d2b09bdc0b59e70299cd03eb578c080155feeb1dd527de0cec62c45e6"} Mar 21 05:45:04 crc kubenswrapper[4775]: I0321 05:45:04.182166 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a47c618d2b09bdc0b59e70299cd03eb578c080155feeb1dd527de0cec62c45e6" Mar 21 05:45:04 crc kubenswrapper[4775]: I0321 05:45:04.181896 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567865-dtkzm" Mar 21 05:45:04 crc kubenswrapper[4775]: I0321 05:45:04.240202 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq"] Mar 21 05:45:04 crc kubenswrapper[4775]: I0321 05:45:04.249546 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567820-pcqhq"] Mar 21 05:45:05 crc kubenswrapper[4775]: I0321 05:45:05.696290 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="421d78dc-59bb-4b5b-9738-1eb6a6144b38" path="/var/lib/kubelet/pods/421d78dc-59bb-4b5b-9738-1eb6a6144b38/volumes" Mar 21 05:45:32 crc kubenswrapper[4775]: I0321 05:45:32.481867 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:45:32 crc kubenswrapper[4775]: I0321 05:45:32.482410 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:45:48 crc kubenswrapper[4775]: I0321 05:45:48.751864 4775 scope.go:117] "RemoveContainer" containerID="19febf6fd2202a1482d38e4d840812756bb31244e9ab9826c5c08685c9d1d70a" Mar 21 05:45:59 crc kubenswrapper[4775]: I0321 05:45:59.973555 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rq7vc"] Mar 21 05:45:59 crc kubenswrapper[4775]: E0321 05:45:59.974468 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="853b684f-d127-48b8-9dd6-5eed4edccfe6" containerName="collect-profiles" Mar 21 05:45:59 crc kubenswrapper[4775]: I0321 05:45:59.974484 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="853b684f-d127-48b8-9dd6-5eed4edccfe6" containerName="collect-profiles" Mar 21 05:45:59 crc kubenswrapper[4775]: I0321 05:45:59.974694 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="853b684f-d127-48b8-9dd6-5eed4edccfe6" containerName="collect-profiles" Mar 21 05:45:59 crc kubenswrapper[4775]: I0321 05:45:59.976231 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rq7vc" Mar 21 05:45:59 crc kubenswrapper[4775]: I0321 05:45:59.996797 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rq7vc"] Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.064718 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad7dafc3-787f-416c-8d24-e14205d49ef9-catalog-content\") pod \"community-operators-rq7vc\" (UID: \"ad7dafc3-787f-416c-8d24-e14205d49ef9\") " pod="openshift-marketplace/community-operators-rq7vc" Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.065104 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htrdd\" (UniqueName: \"kubernetes.io/projected/ad7dafc3-787f-416c-8d24-e14205d49ef9-kube-api-access-htrdd\") pod \"community-operators-rq7vc\" (UID: \"ad7dafc3-787f-416c-8d24-e14205d49ef9\") " pod="openshift-marketplace/community-operators-rq7vc" Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.065223 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad7dafc3-787f-416c-8d24-e14205d49ef9-utilities\") pod \"community-operators-rq7vc\" (UID: \"ad7dafc3-787f-416c-8d24-e14205d49ef9\") " pod="openshift-marketplace/community-operators-rq7vc" Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.148395 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567866-r9qrw"] Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.150076 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567866-r9qrw" Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.154005 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.155226 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.156559 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.159706 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567866-r9qrw"] Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.166740 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htrdd\" (UniqueName: \"kubernetes.io/projected/ad7dafc3-787f-416c-8d24-e14205d49ef9-kube-api-access-htrdd\") pod \"community-operators-rq7vc\" (UID: \"ad7dafc3-787f-416c-8d24-e14205d49ef9\") " pod="openshift-marketplace/community-operators-rq7vc" Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.166814 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad7dafc3-787f-416c-8d24-e14205d49ef9-utilities\") pod \"community-operators-rq7vc\" (UID: \"ad7dafc3-787f-416c-8d24-e14205d49ef9\") " pod="openshift-marketplace/community-operators-rq7vc" Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.166934 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad7dafc3-787f-416c-8d24-e14205d49ef9-catalog-content\") pod \"community-operators-rq7vc\" (UID: \"ad7dafc3-787f-416c-8d24-e14205d49ef9\") " pod="openshift-marketplace/community-operators-rq7vc" Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.167634 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad7dafc3-787f-416c-8d24-e14205d49ef9-catalog-content\") pod \"community-operators-rq7vc\" (UID: \"ad7dafc3-787f-416c-8d24-e14205d49ef9\") " pod="openshift-marketplace/community-operators-rq7vc" Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.167814 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad7dafc3-787f-416c-8d24-e14205d49ef9-utilities\") pod \"community-operators-rq7vc\" (UID: \"ad7dafc3-787f-416c-8d24-e14205d49ef9\") " pod="openshift-marketplace/community-operators-rq7vc" Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.199394 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htrdd\" (UniqueName: \"kubernetes.io/projected/ad7dafc3-787f-416c-8d24-e14205d49ef9-kube-api-access-htrdd\") pod \"community-operators-rq7vc\" (UID: \"ad7dafc3-787f-416c-8d24-e14205d49ef9\") " pod="openshift-marketplace/community-operators-rq7vc" Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.268882 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rncg7\" (UniqueName: \"kubernetes.io/projected/52968d61-f2a8-4bcd-9e97-cb01c508a52e-kube-api-access-rncg7\") pod \"auto-csr-approver-29567866-r9qrw\" (UID: \"52968d61-f2a8-4bcd-9e97-cb01c508a52e\") " pod="openshift-infra/auto-csr-approver-29567866-r9qrw" Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.303331 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rq7vc" Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.370747 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rncg7\" (UniqueName: \"kubernetes.io/projected/52968d61-f2a8-4bcd-9e97-cb01c508a52e-kube-api-access-rncg7\") pod \"auto-csr-approver-29567866-r9qrw\" (UID: \"52968d61-f2a8-4bcd-9e97-cb01c508a52e\") " pod="openshift-infra/auto-csr-approver-29567866-r9qrw" Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.394988 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rncg7\" (UniqueName: \"kubernetes.io/projected/52968d61-f2a8-4bcd-9e97-cb01c508a52e-kube-api-access-rncg7\") pod \"auto-csr-approver-29567866-r9qrw\" (UID: \"52968d61-f2a8-4bcd-9e97-cb01c508a52e\") " pod="openshift-infra/auto-csr-approver-29567866-r9qrw" Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.472786 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567866-r9qrw" Mar 21 05:46:00 crc kubenswrapper[4775]: I0321 05:46:00.916885 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rq7vc"] Mar 21 05:46:01 crc kubenswrapper[4775]: I0321 05:46:01.062608 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567866-r9qrw"] Mar 21 05:46:01 crc kubenswrapper[4775]: W0321 05:46:01.066175 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52968d61_f2a8_4bcd_9e97_cb01c508a52e.slice/crio-80eeb92ee1ea5558f25c003260f13ac5085ae991a234da84daaed0e1fb2960b3 WatchSource:0}: Error finding container 80eeb92ee1ea5558f25c003260f13ac5085ae991a234da84daaed0e1fb2960b3: Status 404 returned error can't find the container with id 80eeb92ee1ea5558f25c003260f13ac5085ae991a234da84daaed0e1fb2960b3 Mar 21 05:46:01 crc kubenswrapper[4775]: I0321 05:46:01.718779 4775 generic.go:334] "Generic (PLEG): container finished" podID="ad7dafc3-787f-416c-8d24-e14205d49ef9" containerID="8dc9d5d632ab88270b647b5baf2266422cef8bce7289c3c70bdb45fd5569924b" exitCode=0 Mar 21 05:46:01 crc kubenswrapper[4775]: I0321 05:46:01.718885 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rq7vc" event={"ID":"ad7dafc3-787f-416c-8d24-e14205d49ef9","Type":"ContainerDied","Data":"8dc9d5d632ab88270b647b5baf2266422cef8bce7289c3c70bdb45fd5569924b"} Mar 21 05:46:01 crc kubenswrapper[4775]: I0321 05:46:01.718959 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rq7vc" event={"ID":"ad7dafc3-787f-416c-8d24-e14205d49ef9","Type":"ContainerStarted","Data":"1a44081b04cb6ba41f124a0f5e6ff0b10d25dc91b013064ce6b7e927c43304be"} Mar 21 05:46:01 crc kubenswrapper[4775]: I0321 05:46:01.720076 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567866-r9qrw" event={"ID":"52968d61-f2a8-4bcd-9e97-cb01c508a52e","Type":"ContainerStarted","Data":"80eeb92ee1ea5558f25c003260f13ac5085ae991a234da84daaed0e1fb2960b3"} Mar 21 05:46:02 crc kubenswrapper[4775]: I0321 05:46:02.482464 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:46:02 crc kubenswrapper[4775]: I0321 05:46:02.483063 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:46:02 crc kubenswrapper[4775]: I0321 05:46:02.483109 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 05:46:02 crc kubenswrapper[4775]: I0321 05:46:02.483906 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"27384562242ad4f001715c32b1c4d680359f25816c963e4f026f07e0e28e85a2"} pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 05:46:02 crc kubenswrapper[4775]: I0321 05:46:02.483956 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" containerID="cri-o://27384562242ad4f001715c32b1c4d680359f25816c963e4f026f07e0e28e85a2" gracePeriod=600 Mar 21 05:46:02 crc kubenswrapper[4775]: I0321 05:46:02.729419 4775 generic.go:334] "Generic (PLEG): container finished" podID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerID="27384562242ad4f001715c32b1c4d680359f25816c963e4f026f07e0e28e85a2" exitCode=0 Mar 21 05:46:02 crc kubenswrapper[4775]: I0321 05:46:02.729497 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerDied","Data":"27384562242ad4f001715c32b1c4d680359f25816c963e4f026f07e0e28e85a2"} Mar 21 05:46:02 crc kubenswrapper[4775]: I0321 05:46:02.729551 4775 scope.go:117] "RemoveContainer" containerID="f0c9b510088c7d635cf52734c112efecde67f87d29c60e9903dd8c0901a91ef9" Mar 21 05:46:02 crc kubenswrapper[4775]: I0321 05:46:02.735669 4775 generic.go:334] "Generic (PLEG): container finished" podID="52968d61-f2a8-4bcd-9e97-cb01c508a52e" containerID="875a527a3f14334ab80b89530523916cfe6a1d4008ed65204aa4a0f63ca04790" exitCode=0 Mar 21 05:46:02 crc kubenswrapper[4775]: I0321 05:46:02.735780 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567866-r9qrw" event={"ID":"52968d61-f2a8-4bcd-9e97-cb01c508a52e","Type":"ContainerDied","Data":"875a527a3f14334ab80b89530523916cfe6a1d4008ed65204aa4a0f63ca04790"} Mar 21 05:46:03 crc kubenswrapper[4775]: I0321 05:46:03.749924 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rq7vc" event={"ID":"ad7dafc3-787f-416c-8d24-e14205d49ef9","Type":"ContainerStarted","Data":"91e8f06f511abc06a57cb2ea3ef15bf4b3dcdd6588f6404ed6c313f7cdc50de8"} Mar 21 05:46:03 crc kubenswrapper[4775]: I0321 05:46:03.752837 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerStarted","Data":"fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2"} Mar 21 05:46:04 crc kubenswrapper[4775]: I0321 05:46:04.176379 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567866-r9qrw" Mar 21 05:46:04 crc kubenswrapper[4775]: I0321 05:46:04.270882 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rncg7\" (UniqueName: \"kubernetes.io/projected/52968d61-f2a8-4bcd-9e97-cb01c508a52e-kube-api-access-rncg7\") pod \"52968d61-f2a8-4bcd-9e97-cb01c508a52e\" (UID: \"52968d61-f2a8-4bcd-9e97-cb01c508a52e\") " Mar 21 05:46:04 crc kubenswrapper[4775]: I0321 05:46:04.277623 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52968d61-f2a8-4bcd-9e97-cb01c508a52e-kube-api-access-rncg7" (OuterVolumeSpecName: "kube-api-access-rncg7") pod "52968d61-f2a8-4bcd-9e97-cb01c508a52e" (UID: "52968d61-f2a8-4bcd-9e97-cb01c508a52e"). InnerVolumeSpecName "kube-api-access-rncg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:46:04 crc kubenswrapper[4775]: I0321 05:46:04.373074 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rncg7\" (UniqueName: \"kubernetes.io/projected/52968d61-f2a8-4bcd-9e97-cb01c508a52e-kube-api-access-rncg7\") on node \"crc\" DevicePath \"\"" Mar 21 05:46:04 crc kubenswrapper[4775]: I0321 05:46:04.762813 4775 generic.go:334] "Generic (PLEG): container finished" podID="ad7dafc3-787f-416c-8d24-e14205d49ef9" containerID="91e8f06f511abc06a57cb2ea3ef15bf4b3dcdd6588f6404ed6c313f7cdc50de8" exitCode=0 Mar 21 05:46:04 crc kubenswrapper[4775]: I0321 05:46:04.762861 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rq7vc" event={"ID":"ad7dafc3-787f-416c-8d24-e14205d49ef9","Type":"ContainerDied","Data":"91e8f06f511abc06a57cb2ea3ef15bf4b3dcdd6588f6404ed6c313f7cdc50de8"} Mar 21 05:46:04 crc kubenswrapper[4775]: I0321 05:46:04.764676 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567866-r9qrw" Mar 21 05:46:04 crc kubenswrapper[4775]: I0321 05:46:04.764648 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567866-r9qrw" event={"ID":"52968d61-f2a8-4bcd-9e97-cb01c508a52e","Type":"ContainerDied","Data":"80eeb92ee1ea5558f25c003260f13ac5085ae991a234da84daaed0e1fb2960b3"} Mar 21 05:46:04 crc kubenswrapper[4775]: I0321 05:46:04.764741 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80eeb92ee1ea5558f25c003260f13ac5085ae991a234da84daaed0e1fb2960b3" Mar 21 05:46:05 crc kubenswrapper[4775]: I0321 05:46:05.247151 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567860-f4gsq"] Mar 21 05:46:05 crc kubenswrapper[4775]: I0321 05:46:05.257547 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567860-f4gsq"] Mar 21 05:46:05 crc kubenswrapper[4775]: I0321 05:46:05.672331 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42fbc24f-3528-4778-ae85-23e0c2fe99d9" path="/var/lib/kubelet/pods/42fbc24f-3528-4778-ae85-23e0c2fe99d9/volumes" Mar 21 05:46:05 crc kubenswrapper[4775]: I0321 05:46:05.776202 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rq7vc" event={"ID":"ad7dafc3-787f-416c-8d24-e14205d49ef9","Type":"ContainerStarted","Data":"05a691707ef3970b49493cbe9973459894a5f44eb36bfdb928a95355b3928366"} Mar 21 05:46:05 crc kubenswrapper[4775]: I0321 05:46:05.794704 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rq7vc" podStartSLOduration=3.294294484 podStartE2EDuration="6.794690562s" podCreationTimestamp="2026-03-21 05:45:59 +0000 UTC" firstStartedPulling="2026-03-21 05:46:01.720686624 +0000 UTC m=+3514.697150248" lastFinishedPulling="2026-03-21 05:46:05.221082702 +0000 UTC m=+3518.197546326" observedRunningTime="2026-03-21 05:46:05.794091125 +0000 UTC m=+3518.770554749" watchObservedRunningTime="2026-03-21 05:46:05.794690562 +0000 UTC m=+3518.771154186" Mar 21 05:46:10 crc kubenswrapper[4775]: I0321 05:46:10.304703 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rq7vc" Mar 21 05:46:10 crc kubenswrapper[4775]: I0321 05:46:10.306669 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rq7vc" Mar 21 05:46:10 crc kubenswrapper[4775]: I0321 05:46:10.411347 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rq7vc" Mar 21 05:46:10 crc kubenswrapper[4775]: I0321 05:46:10.888316 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rq7vc" Mar 21 05:46:10 crc kubenswrapper[4775]: I0321 05:46:10.942735 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rq7vc"] Mar 21 05:46:12 crc kubenswrapper[4775]: I0321 05:46:12.844353 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rq7vc" podUID="ad7dafc3-787f-416c-8d24-e14205d49ef9" containerName="registry-server" containerID="cri-o://05a691707ef3970b49493cbe9973459894a5f44eb36bfdb928a95355b3928366" gracePeriod=2 Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.373005 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rq7vc" Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.453526 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad7dafc3-787f-416c-8d24-e14205d49ef9-catalog-content\") pod \"ad7dafc3-787f-416c-8d24-e14205d49ef9\" (UID: \"ad7dafc3-787f-416c-8d24-e14205d49ef9\") " Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.453576 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htrdd\" (UniqueName: \"kubernetes.io/projected/ad7dafc3-787f-416c-8d24-e14205d49ef9-kube-api-access-htrdd\") pod \"ad7dafc3-787f-416c-8d24-e14205d49ef9\" (UID: \"ad7dafc3-787f-416c-8d24-e14205d49ef9\") " Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.453657 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad7dafc3-787f-416c-8d24-e14205d49ef9-utilities\") pod \"ad7dafc3-787f-416c-8d24-e14205d49ef9\" (UID: \"ad7dafc3-787f-416c-8d24-e14205d49ef9\") " Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.454771 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad7dafc3-787f-416c-8d24-e14205d49ef9-utilities" (OuterVolumeSpecName: "utilities") pod "ad7dafc3-787f-416c-8d24-e14205d49ef9" (UID: "ad7dafc3-787f-416c-8d24-e14205d49ef9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.460192 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad7dafc3-787f-416c-8d24-e14205d49ef9-kube-api-access-htrdd" (OuterVolumeSpecName: "kube-api-access-htrdd") pod "ad7dafc3-787f-416c-8d24-e14205d49ef9" (UID: "ad7dafc3-787f-416c-8d24-e14205d49ef9"). InnerVolumeSpecName "kube-api-access-htrdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.520991 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad7dafc3-787f-416c-8d24-e14205d49ef9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad7dafc3-787f-416c-8d24-e14205d49ef9" (UID: "ad7dafc3-787f-416c-8d24-e14205d49ef9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.556327 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad7dafc3-787f-416c-8d24-e14205d49ef9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.556367 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htrdd\" (UniqueName: \"kubernetes.io/projected/ad7dafc3-787f-416c-8d24-e14205d49ef9-kube-api-access-htrdd\") on node \"crc\" DevicePath \"\"" Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.556376 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad7dafc3-787f-416c-8d24-e14205d49ef9-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.855490 4775 generic.go:334] "Generic (PLEG): container finished" podID="ad7dafc3-787f-416c-8d24-e14205d49ef9" containerID="05a691707ef3970b49493cbe9973459894a5f44eb36bfdb928a95355b3928366" exitCode=0 Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.855588 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rq7vc" event={"ID":"ad7dafc3-787f-416c-8d24-e14205d49ef9","Type":"ContainerDied","Data":"05a691707ef3970b49493cbe9973459894a5f44eb36bfdb928a95355b3928366"} Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.855880 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rq7vc" event={"ID":"ad7dafc3-787f-416c-8d24-e14205d49ef9","Type":"ContainerDied","Data":"1a44081b04cb6ba41f124a0f5e6ff0b10d25dc91b013064ce6b7e927c43304be"} Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.855905 4775 scope.go:117] "RemoveContainer" containerID="05a691707ef3970b49493cbe9973459894a5f44eb36bfdb928a95355b3928366" Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.855595 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rq7vc" Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.884087 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rq7vc"] Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.886710 4775 scope.go:117] "RemoveContainer" containerID="91e8f06f511abc06a57cb2ea3ef15bf4b3dcdd6588f6404ed6c313f7cdc50de8" Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.894889 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rq7vc"] Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.911525 4775 scope.go:117] "RemoveContainer" containerID="8dc9d5d632ab88270b647b5baf2266422cef8bce7289c3c70bdb45fd5569924b" Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.963752 4775 scope.go:117] "RemoveContainer" containerID="05a691707ef3970b49493cbe9973459894a5f44eb36bfdb928a95355b3928366" Mar 21 05:46:13 crc kubenswrapper[4775]: E0321 05:46:13.964355 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05a691707ef3970b49493cbe9973459894a5f44eb36bfdb928a95355b3928366\": container with ID starting with 05a691707ef3970b49493cbe9973459894a5f44eb36bfdb928a95355b3928366 not found: ID does not exist" containerID="05a691707ef3970b49493cbe9973459894a5f44eb36bfdb928a95355b3928366" Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.964425 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05a691707ef3970b49493cbe9973459894a5f44eb36bfdb928a95355b3928366"} err="failed to get container status \"05a691707ef3970b49493cbe9973459894a5f44eb36bfdb928a95355b3928366\": rpc error: code = NotFound desc = could not find container \"05a691707ef3970b49493cbe9973459894a5f44eb36bfdb928a95355b3928366\": container with ID starting with 05a691707ef3970b49493cbe9973459894a5f44eb36bfdb928a95355b3928366 not found: ID does not exist" Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.964464 4775 scope.go:117] "RemoveContainer" containerID="91e8f06f511abc06a57cb2ea3ef15bf4b3dcdd6588f6404ed6c313f7cdc50de8" Mar 21 05:46:13 crc kubenswrapper[4775]: E0321 05:46:13.964802 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91e8f06f511abc06a57cb2ea3ef15bf4b3dcdd6588f6404ed6c313f7cdc50de8\": container with ID starting with 91e8f06f511abc06a57cb2ea3ef15bf4b3dcdd6588f6404ed6c313f7cdc50de8 not found: ID does not exist" containerID="91e8f06f511abc06a57cb2ea3ef15bf4b3dcdd6588f6404ed6c313f7cdc50de8" Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.964835 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91e8f06f511abc06a57cb2ea3ef15bf4b3dcdd6588f6404ed6c313f7cdc50de8"} err="failed to get container status \"91e8f06f511abc06a57cb2ea3ef15bf4b3dcdd6588f6404ed6c313f7cdc50de8\": rpc error: code = NotFound desc = could not find container \"91e8f06f511abc06a57cb2ea3ef15bf4b3dcdd6588f6404ed6c313f7cdc50de8\": container with ID starting with 91e8f06f511abc06a57cb2ea3ef15bf4b3dcdd6588f6404ed6c313f7cdc50de8 not found: ID does not exist" Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.964857 4775 scope.go:117] "RemoveContainer" containerID="8dc9d5d632ab88270b647b5baf2266422cef8bce7289c3c70bdb45fd5569924b" Mar 21 05:46:13 crc kubenswrapper[4775]: E0321 05:46:13.965378 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dc9d5d632ab88270b647b5baf2266422cef8bce7289c3c70bdb45fd5569924b\": container with ID starting with 8dc9d5d632ab88270b647b5baf2266422cef8bce7289c3c70bdb45fd5569924b not found: ID does not exist" containerID="8dc9d5d632ab88270b647b5baf2266422cef8bce7289c3c70bdb45fd5569924b" Mar 21 05:46:13 crc kubenswrapper[4775]: I0321 05:46:13.965454 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dc9d5d632ab88270b647b5baf2266422cef8bce7289c3c70bdb45fd5569924b"} err="failed to get container status \"8dc9d5d632ab88270b647b5baf2266422cef8bce7289c3c70bdb45fd5569924b\": rpc error: code = NotFound desc = could not find container \"8dc9d5d632ab88270b647b5baf2266422cef8bce7289c3c70bdb45fd5569924b\": container with ID starting with 8dc9d5d632ab88270b647b5baf2266422cef8bce7289c3c70bdb45fd5569924b not found: ID does not exist" Mar 21 05:46:15 crc kubenswrapper[4775]: I0321 05:46:15.672799 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad7dafc3-787f-416c-8d24-e14205d49ef9" path="/var/lib/kubelet/pods/ad7dafc3-787f-416c-8d24-e14205d49ef9/volumes" Mar 21 05:46:24 crc kubenswrapper[4775]: I0321 05:46:24.981695 4775 generic.go:334] "Generic (PLEG): container finished" podID="1c832898-838d-423d-8ad8-512c5ee5706c" containerID="e5d5a6749ca728ee30628ed8c23d0df8c8193d878227278bf224646abd9c8806" exitCode=0 Mar 21 05:46:24 crc kubenswrapper[4775]: I0321 05:46:24.982287 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1c832898-838d-423d-8ad8-512c5ee5706c","Type":"ContainerDied","Data":"e5d5a6749ca728ee30628ed8c23d0df8c8193d878227278bf224646abd9c8806"} Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.377985 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.505963 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1c832898-838d-423d-8ad8-512c5ee5706c-config-data\") pod \"1c832898-838d-423d-8ad8-512c5ee5706c\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.506129 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1c832898-838d-423d-8ad8-512c5ee5706c-ca-certs\") pod \"1c832898-838d-423d-8ad8-512c5ee5706c\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.506206 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"1c832898-838d-423d-8ad8-512c5ee5706c\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.506243 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1c832898-838d-423d-8ad8-512c5ee5706c-test-operator-ephemeral-workdir\") pod \"1c832898-838d-423d-8ad8-512c5ee5706c\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.506289 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1c832898-838d-423d-8ad8-512c5ee5706c-test-operator-ephemeral-temporary\") pod \"1c832898-838d-423d-8ad8-512c5ee5706c\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.506308 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1c832898-838d-423d-8ad8-512c5ee5706c-openstack-config-secret\") pod \"1c832898-838d-423d-8ad8-512c5ee5706c\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.506385 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdw2v\" (UniqueName: \"kubernetes.io/projected/1c832898-838d-423d-8ad8-512c5ee5706c-kube-api-access-gdw2v\") pod \"1c832898-838d-423d-8ad8-512c5ee5706c\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.506444 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1c832898-838d-423d-8ad8-512c5ee5706c-openstack-config\") pod \"1c832898-838d-423d-8ad8-512c5ee5706c\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.506507 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1c832898-838d-423d-8ad8-512c5ee5706c-ssh-key\") pod \"1c832898-838d-423d-8ad8-512c5ee5706c\" (UID: \"1c832898-838d-423d-8ad8-512c5ee5706c\") " Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.507030 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c832898-838d-423d-8ad8-512c5ee5706c-config-data" (OuterVolumeSpecName: "config-data") pod "1c832898-838d-423d-8ad8-512c5ee5706c" (UID: "1c832898-838d-423d-8ad8-512c5ee5706c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.507647 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c832898-838d-423d-8ad8-512c5ee5706c-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "1c832898-838d-423d-8ad8-512c5ee5706c" (UID: "1c832898-838d-423d-8ad8-512c5ee5706c"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.512085 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "test-operator-logs") pod "1c832898-838d-423d-8ad8-512c5ee5706c" (UID: "1c832898-838d-423d-8ad8-512c5ee5706c"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.512549 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c832898-838d-423d-8ad8-512c5ee5706c-kube-api-access-gdw2v" (OuterVolumeSpecName: "kube-api-access-gdw2v") pod "1c832898-838d-423d-8ad8-512c5ee5706c" (UID: "1c832898-838d-423d-8ad8-512c5ee5706c"). InnerVolumeSpecName "kube-api-access-gdw2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.517389 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c832898-838d-423d-8ad8-512c5ee5706c-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "1c832898-838d-423d-8ad8-512c5ee5706c" (UID: "1c832898-838d-423d-8ad8-512c5ee5706c"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.536010 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c832898-838d-423d-8ad8-512c5ee5706c-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "1c832898-838d-423d-8ad8-512c5ee5706c" (UID: "1c832898-838d-423d-8ad8-512c5ee5706c"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.540148 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c832898-838d-423d-8ad8-512c5ee5706c-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "1c832898-838d-423d-8ad8-512c5ee5706c" (UID: "1c832898-838d-423d-8ad8-512c5ee5706c"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.564474 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c832898-838d-423d-8ad8-512c5ee5706c-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "1c832898-838d-423d-8ad8-512c5ee5706c" (UID: "1c832898-838d-423d-8ad8-512c5ee5706c"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.564982 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c832898-838d-423d-8ad8-512c5ee5706c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1c832898-838d-423d-8ad8-512c5ee5706c" (UID: "1c832898-838d-423d-8ad8-512c5ee5706c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.608461 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1c832898-838d-423d-8ad8-512c5ee5706c-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.608513 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1c832898-838d-423d-8ad8-512c5ee5706c-ssh-key\") on node \"crc\" DevicePath \"\"" Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.608522 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1c832898-838d-423d-8ad8-512c5ee5706c-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.608531 4775 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1c832898-838d-423d-8ad8-512c5ee5706c-ca-certs\") on node \"crc\" DevicePath \"\"" Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.608566 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.608577 4775 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1c832898-838d-423d-8ad8-512c5ee5706c-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.608586 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1c832898-838d-423d-8ad8-512c5ee5706c-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.608596 4775 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1c832898-838d-423d-8ad8-512c5ee5706c-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.608605 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdw2v\" (UniqueName: \"kubernetes.io/projected/1c832898-838d-423d-8ad8-512c5ee5706c-kube-api-access-gdw2v\") on node \"crc\" DevicePath \"\"" Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.627205 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Mar 21 05:46:26 crc kubenswrapper[4775]: I0321 05:46:26.711147 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Mar 21 05:46:27 crc kubenswrapper[4775]: I0321 05:46:27.000349 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1c832898-838d-423d-8ad8-512c5ee5706c","Type":"ContainerDied","Data":"695de001f7207ca22144c3ce97e3ea6306a89b43020987a1d4856f3f9e289416"} Mar 21 05:46:27 crc kubenswrapper[4775]: I0321 05:46:27.000407 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="695de001f7207ca22144c3ce97e3ea6306a89b43020987a1d4856f3f9e289416" Mar 21 05:46:27 crc kubenswrapper[4775]: I0321 05:46:27.000458 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 21 05:46:33 crc kubenswrapper[4775]: I0321 05:46:33.875724 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 21 05:46:33 crc kubenswrapper[4775]: E0321 05:46:33.876636 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad7dafc3-787f-416c-8d24-e14205d49ef9" containerName="extract-content" Mar 21 05:46:33 crc kubenswrapper[4775]: I0321 05:46:33.876686 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad7dafc3-787f-416c-8d24-e14205d49ef9" containerName="extract-content" Mar 21 05:46:33 crc kubenswrapper[4775]: E0321 05:46:33.876722 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad7dafc3-787f-416c-8d24-e14205d49ef9" containerName="registry-server" Mar 21 05:46:33 crc kubenswrapper[4775]: I0321 05:46:33.876732 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad7dafc3-787f-416c-8d24-e14205d49ef9" containerName="registry-server" Mar 21 05:46:33 crc kubenswrapper[4775]: E0321 05:46:33.876764 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c832898-838d-423d-8ad8-512c5ee5706c" containerName="tempest-tests-tempest-tests-runner" Mar 21 05:46:33 crc kubenswrapper[4775]: I0321 05:46:33.876774 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c832898-838d-423d-8ad8-512c5ee5706c" containerName="tempest-tests-tempest-tests-runner" Mar 21 05:46:33 crc kubenswrapper[4775]: E0321 05:46:33.876792 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad7dafc3-787f-416c-8d24-e14205d49ef9" containerName="extract-utilities" Mar 21 05:46:33 crc kubenswrapper[4775]: I0321 05:46:33.876802 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad7dafc3-787f-416c-8d24-e14205d49ef9" containerName="extract-utilities" Mar 21 05:46:33 crc kubenswrapper[4775]: E0321 05:46:33.876819 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52968d61-f2a8-4bcd-9e97-cb01c508a52e" containerName="oc" Mar 21 05:46:33 crc kubenswrapper[4775]: I0321 05:46:33.876828 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="52968d61-f2a8-4bcd-9e97-cb01c508a52e" containerName="oc" Mar 21 05:46:33 crc kubenswrapper[4775]: I0321 05:46:33.877085 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="52968d61-f2a8-4bcd-9e97-cb01c508a52e" containerName="oc" Mar 21 05:46:33 crc kubenswrapper[4775]: I0321 05:46:33.877142 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad7dafc3-787f-416c-8d24-e14205d49ef9" containerName="registry-server" Mar 21 05:46:33 crc kubenswrapper[4775]: I0321 05:46:33.877160 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c832898-838d-423d-8ad8-512c5ee5706c" containerName="tempest-tests-tempest-tests-runner" Mar 21 05:46:33 crc kubenswrapper[4775]: I0321 05:46:33.878011 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 21 05:46:33 crc kubenswrapper[4775]: I0321 05:46:33.881100 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-bnv6q" Mar 21 05:46:33 crc kubenswrapper[4775]: I0321 05:46:33.889592 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 21 05:46:33 crc kubenswrapper[4775]: I0321 05:46:33.965996 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74n4l\" (UniqueName: \"kubernetes.io/projected/2c99faa4-db71-4a05-a018-9c382f33f55e-kube-api-access-74n4l\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2c99faa4-db71-4a05-a018-9c382f33f55e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 21 05:46:33 crc kubenswrapper[4775]: I0321 05:46:33.966428 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2c99faa4-db71-4a05-a018-9c382f33f55e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 21 05:46:34 crc kubenswrapper[4775]: I0321 05:46:34.068922 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74n4l\" (UniqueName: \"kubernetes.io/projected/2c99faa4-db71-4a05-a018-9c382f33f55e-kube-api-access-74n4l\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2c99faa4-db71-4a05-a018-9c382f33f55e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 21 05:46:34 crc kubenswrapper[4775]: I0321 05:46:34.069043 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2c99faa4-db71-4a05-a018-9c382f33f55e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 21 05:46:34 crc kubenswrapper[4775]: I0321 05:46:34.069568 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2c99faa4-db71-4a05-a018-9c382f33f55e\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 21 05:46:34 crc kubenswrapper[4775]: I0321 05:46:34.092836 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74n4l\" (UniqueName: \"kubernetes.io/projected/2c99faa4-db71-4a05-a018-9c382f33f55e-kube-api-access-74n4l\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2c99faa4-db71-4a05-a018-9c382f33f55e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 21 05:46:34 crc kubenswrapper[4775]: I0321 05:46:34.101171 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2c99faa4-db71-4a05-a018-9c382f33f55e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 21 05:46:34 crc kubenswrapper[4775]: I0321 05:46:34.210873 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 21 05:46:34 crc kubenswrapper[4775]: I0321 05:46:34.650412 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 21 05:46:35 crc kubenswrapper[4775]: I0321 05:46:35.105690 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"2c99faa4-db71-4a05-a018-9c382f33f55e","Type":"ContainerStarted","Data":"a1b6e5eb529eddaf4e17d335f3dd4d0c4c8307feaa5af921a19809433863ab62"} Mar 21 05:46:36 crc kubenswrapper[4775]: I0321 05:46:36.115911 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"2c99faa4-db71-4a05-a018-9c382f33f55e","Type":"ContainerStarted","Data":"a379f3260971c9b4bf0cc3e889ee5a76416e8393c60c47aa6b68e112ce28ff1e"} Mar 21 05:46:36 crc kubenswrapper[4775]: I0321 05:46:36.132980 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.314587789 podStartE2EDuration="3.13296503s" podCreationTimestamp="2026-03-21 05:46:33 +0000 UTC" firstStartedPulling="2026-03-21 05:46:34.65304146 +0000 UTC m=+3547.629505164" lastFinishedPulling="2026-03-21 05:46:35.471418781 +0000 UTC m=+3548.447882405" observedRunningTime="2026-03-21 05:46:36.130790128 +0000 UTC m=+3549.107253762" watchObservedRunningTime="2026-03-21 05:46:36.13296503 +0000 UTC m=+3549.109428654" Mar 21 05:46:47 crc kubenswrapper[4775]: I0321 05:46:47.061891 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mtx69"] Mar 21 05:46:47 crc kubenswrapper[4775]: I0321 05:46:47.068674 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mtx69" Mar 21 05:46:47 crc kubenswrapper[4775]: I0321 05:46:47.074950 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mtx69"] Mar 21 05:46:47 crc kubenswrapper[4775]: I0321 05:46:47.186680 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fa06424-b2f2-4965-85d6-52f7824c1663-utilities\") pod \"redhat-marketplace-mtx69\" (UID: \"7fa06424-b2f2-4965-85d6-52f7824c1663\") " pod="openshift-marketplace/redhat-marketplace-mtx69" Mar 21 05:46:47 crc kubenswrapper[4775]: I0321 05:46:47.186750 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfcts\" (UniqueName: \"kubernetes.io/projected/7fa06424-b2f2-4965-85d6-52f7824c1663-kube-api-access-rfcts\") pod \"redhat-marketplace-mtx69\" (UID: \"7fa06424-b2f2-4965-85d6-52f7824c1663\") " pod="openshift-marketplace/redhat-marketplace-mtx69" Mar 21 05:46:47 crc kubenswrapper[4775]: I0321 05:46:47.186842 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fa06424-b2f2-4965-85d6-52f7824c1663-catalog-content\") pod \"redhat-marketplace-mtx69\" (UID: \"7fa06424-b2f2-4965-85d6-52f7824c1663\") " pod="openshift-marketplace/redhat-marketplace-mtx69" Mar 21 05:46:47 crc kubenswrapper[4775]: I0321 05:46:47.288685 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fa06424-b2f2-4965-85d6-52f7824c1663-utilities\") pod \"redhat-marketplace-mtx69\" (UID: \"7fa06424-b2f2-4965-85d6-52f7824c1663\") " pod="openshift-marketplace/redhat-marketplace-mtx69" Mar 21 05:46:47 crc kubenswrapper[4775]: I0321 05:46:47.288763 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfcts\" (UniqueName: \"kubernetes.io/projected/7fa06424-b2f2-4965-85d6-52f7824c1663-kube-api-access-rfcts\") pod \"redhat-marketplace-mtx69\" (UID: \"7fa06424-b2f2-4965-85d6-52f7824c1663\") " pod="openshift-marketplace/redhat-marketplace-mtx69" Mar 21 05:46:47 crc kubenswrapper[4775]: I0321 05:46:47.288878 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fa06424-b2f2-4965-85d6-52f7824c1663-catalog-content\") pod \"redhat-marketplace-mtx69\" (UID: \"7fa06424-b2f2-4965-85d6-52f7824c1663\") " pod="openshift-marketplace/redhat-marketplace-mtx69" Mar 21 05:46:47 crc kubenswrapper[4775]: I0321 05:46:47.289456 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fa06424-b2f2-4965-85d6-52f7824c1663-catalog-content\") pod \"redhat-marketplace-mtx69\" (UID: \"7fa06424-b2f2-4965-85d6-52f7824c1663\") " pod="openshift-marketplace/redhat-marketplace-mtx69" Mar 21 05:46:47 crc kubenswrapper[4775]: I0321 05:46:47.289495 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fa06424-b2f2-4965-85d6-52f7824c1663-utilities\") pod \"redhat-marketplace-mtx69\" (UID: \"7fa06424-b2f2-4965-85d6-52f7824c1663\") " pod="openshift-marketplace/redhat-marketplace-mtx69" Mar 21 05:46:47 crc kubenswrapper[4775]: I0321 05:46:47.313654 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfcts\" (UniqueName: \"kubernetes.io/projected/7fa06424-b2f2-4965-85d6-52f7824c1663-kube-api-access-rfcts\") pod \"redhat-marketplace-mtx69\" (UID: \"7fa06424-b2f2-4965-85d6-52f7824c1663\") " pod="openshift-marketplace/redhat-marketplace-mtx69" Mar 21 05:46:47 crc kubenswrapper[4775]: I0321 05:46:47.407578 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mtx69" Mar 21 05:46:47 crc kubenswrapper[4775]: I0321 05:46:47.917756 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mtx69"] Mar 21 05:46:47 crc kubenswrapper[4775]: W0321 05:46:47.932138 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fa06424_b2f2_4965_85d6_52f7824c1663.slice/crio-adafd4efc773e168282a543dba92bbe9e306c86f368448a6d3525650dc94f58a WatchSource:0}: Error finding container adafd4efc773e168282a543dba92bbe9e306c86f368448a6d3525650dc94f58a: Status 404 returned error can't find the container with id adafd4efc773e168282a543dba92bbe9e306c86f368448a6d3525650dc94f58a Mar 21 05:46:48 crc kubenswrapper[4775]: I0321 05:46:48.250061 4775 generic.go:334] "Generic (PLEG): container finished" podID="7fa06424-b2f2-4965-85d6-52f7824c1663" containerID="8c9a5a3c4fe60a9ba70c2bdb23fc13eb6fbc93c9d639fb16cff54b7b76bd3b9c" exitCode=0 Mar 21 05:46:48 crc kubenswrapper[4775]: I0321 05:46:48.250154 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtx69" event={"ID":"7fa06424-b2f2-4965-85d6-52f7824c1663","Type":"ContainerDied","Data":"8c9a5a3c4fe60a9ba70c2bdb23fc13eb6fbc93c9d639fb16cff54b7b76bd3b9c"} Mar 21 05:46:48 crc kubenswrapper[4775]: I0321 05:46:48.250213 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtx69" event={"ID":"7fa06424-b2f2-4965-85d6-52f7824c1663","Type":"ContainerStarted","Data":"adafd4efc773e168282a543dba92bbe9e306c86f368448a6d3525650dc94f58a"} Mar 21 05:46:48 crc kubenswrapper[4775]: I0321 05:46:48.806425 4775 scope.go:117] "RemoveContainer" containerID="772f50d7983f92c55a24cd77459f228985481da63abdb797f0dc2f295541ac61" Mar 21 05:46:49 crc kubenswrapper[4775]: I0321 05:46:49.267775 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtx69" event={"ID":"7fa06424-b2f2-4965-85d6-52f7824c1663","Type":"ContainerStarted","Data":"d0545665cdd8e41c947743dcd9829bdc046d8dd1683e708f87ff1c576ef1ba0f"} Mar 21 05:46:50 crc kubenswrapper[4775]: I0321 05:46:50.279202 4775 generic.go:334] "Generic (PLEG): container finished" podID="7fa06424-b2f2-4965-85d6-52f7824c1663" containerID="d0545665cdd8e41c947743dcd9829bdc046d8dd1683e708f87ff1c576ef1ba0f" exitCode=0 Mar 21 05:46:50 crc kubenswrapper[4775]: I0321 05:46:50.279316 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtx69" event={"ID":"7fa06424-b2f2-4965-85d6-52f7824c1663","Type":"ContainerDied","Data":"d0545665cdd8e41c947743dcd9829bdc046d8dd1683e708f87ff1c576ef1ba0f"} Mar 21 05:46:50 crc kubenswrapper[4775]: I0321 05:46:50.279757 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtx69" event={"ID":"7fa06424-b2f2-4965-85d6-52f7824c1663","Type":"ContainerStarted","Data":"82a27cb328425f943a3f5240bbcae1d99975930a49e7720038822aa648d7b9d5"} Mar 21 05:46:50 crc kubenswrapper[4775]: I0321 05:46:50.317490 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mtx69" podStartSLOduration=1.883296724 podStartE2EDuration="3.31747054s" podCreationTimestamp="2026-03-21 05:46:47 +0000 UTC" firstStartedPulling="2026-03-21 05:46:48.25168906 +0000 UTC m=+3561.228152684" lastFinishedPulling="2026-03-21 05:46:49.685862876 +0000 UTC m=+3562.662326500" observedRunningTime="2026-03-21 05:46:50.311085528 +0000 UTC m=+3563.287549162" watchObservedRunningTime="2026-03-21 05:46:50.31747054 +0000 UTC m=+3563.293934164" Mar 21 05:46:57 crc kubenswrapper[4775]: I0321 05:46:57.407974 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mtx69" Mar 21 05:46:57 crc kubenswrapper[4775]: I0321 05:46:57.408481 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mtx69" Mar 21 05:46:57 crc kubenswrapper[4775]: I0321 05:46:57.456079 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mtx69" Mar 21 05:46:58 crc kubenswrapper[4775]: I0321 05:46:58.398107 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mtx69" Mar 21 05:47:01 crc kubenswrapper[4775]: I0321 05:47:01.025025 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mtx69"] Mar 21 05:47:01 crc kubenswrapper[4775]: I0321 05:47:01.025588 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mtx69" podUID="7fa06424-b2f2-4965-85d6-52f7824c1663" containerName="registry-server" containerID="cri-o://82a27cb328425f943a3f5240bbcae1d99975930a49e7720038822aa648d7b9d5" gracePeriod=2 Mar 21 05:47:01 crc kubenswrapper[4775]: I0321 05:47:01.381738 4775 generic.go:334] "Generic (PLEG): container finished" podID="7fa06424-b2f2-4965-85d6-52f7824c1663" containerID="82a27cb328425f943a3f5240bbcae1d99975930a49e7720038822aa648d7b9d5" exitCode=0 Mar 21 05:47:01 crc kubenswrapper[4775]: I0321 05:47:01.381781 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtx69" event={"ID":"7fa06424-b2f2-4965-85d6-52f7824c1663","Type":"ContainerDied","Data":"82a27cb328425f943a3f5240bbcae1d99975930a49e7720038822aa648d7b9d5"} Mar 21 05:47:02 crc kubenswrapper[4775]: I0321 05:47:02.098254 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mtx69" Mar 21 05:47:02 crc kubenswrapper[4775]: I0321 05:47:02.286489 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfcts\" (UniqueName: \"kubernetes.io/projected/7fa06424-b2f2-4965-85d6-52f7824c1663-kube-api-access-rfcts\") pod \"7fa06424-b2f2-4965-85d6-52f7824c1663\" (UID: \"7fa06424-b2f2-4965-85d6-52f7824c1663\") " Mar 21 05:47:02 crc kubenswrapper[4775]: I0321 05:47:02.286838 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fa06424-b2f2-4965-85d6-52f7824c1663-catalog-content\") pod \"7fa06424-b2f2-4965-85d6-52f7824c1663\" (UID: \"7fa06424-b2f2-4965-85d6-52f7824c1663\") " Mar 21 05:47:02 crc kubenswrapper[4775]: I0321 05:47:02.287109 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fa06424-b2f2-4965-85d6-52f7824c1663-utilities\") pod \"7fa06424-b2f2-4965-85d6-52f7824c1663\" (UID: \"7fa06424-b2f2-4965-85d6-52f7824c1663\") " Mar 21 05:47:02 crc kubenswrapper[4775]: I0321 05:47:02.287828 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fa06424-b2f2-4965-85d6-52f7824c1663-utilities" (OuterVolumeSpecName: "utilities") pod "7fa06424-b2f2-4965-85d6-52f7824c1663" (UID: "7fa06424-b2f2-4965-85d6-52f7824c1663"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:47:02 crc kubenswrapper[4775]: I0321 05:47:02.288091 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fa06424-b2f2-4965-85d6-52f7824c1663-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:47:02 crc kubenswrapper[4775]: I0321 05:47:02.292286 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fa06424-b2f2-4965-85d6-52f7824c1663-kube-api-access-rfcts" (OuterVolumeSpecName: "kube-api-access-rfcts") pod "7fa06424-b2f2-4965-85d6-52f7824c1663" (UID: "7fa06424-b2f2-4965-85d6-52f7824c1663"). InnerVolumeSpecName "kube-api-access-rfcts". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:47:02 crc kubenswrapper[4775]: I0321 05:47:02.317411 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fa06424-b2f2-4965-85d6-52f7824c1663-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fa06424-b2f2-4965-85d6-52f7824c1663" (UID: "7fa06424-b2f2-4965-85d6-52f7824c1663"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:47:02 crc kubenswrapper[4775]: I0321 05:47:02.403201 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfcts\" (UniqueName: \"kubernetes.io/projected/7fa06424-b2f2-4965-85d6-52f7824c1663-kube-api-access-rfcts\") on node \"crc\" DevicePath \"\"" Mar 21 05:47:02 crc kubenswrapper[4775]: I0321 05:47:02.403247 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fa06424-b2f2-4965-85d6-52f7824c1663-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:47:02 crc kubenswrapper[4775]: I0321 05:47:02.408425 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtx69" event={"ID":"7fa06424-b2f2-4965-85d6-52f7824c1663","Type":"ContainerDied","Data":"adafd4efc773e168282a543dba92bbe9e306c86f368448a6d3525650dc94f58a"} Mar 21 05:47:02 crc kubenswrapper[4775]: I0321 05:47:02.408486 4775 scope.go:117] "RemoveContainer" containerID="82a27cb328425f943a3f5240bbcae1d99975930a49e7720038822aa648d7b9d5" Mar 21 05:47:02 crc kubenswrapper[4775]: I0321 05:47:02.408613 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mtx69" Mar 21 05:47:02 crc kubenswrapper[4775]: I0321 05:47:02.437460 4775 scope.go:117] "RemoveContainer" containerID="d0545665cdd8e41c947743dcd9829bdc046d8dd1683e708f87ff1c576ef1ba0f" Mar 21 05:47:02 crc kubenswrapper[4775]: I0321 05:47:02.448157 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mtx69"] Mar 21 05:47:02 crc kubenswrapper[4775]: I0321 05:47:02.455307 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mtx69"] Mar 21 05:47:02 crc kubenswrapper[4775]: I0321 05:47:02.458817 4775 scope.go:117] "RemoveContainer" containerID="8c9a5a3c4fe60a9ba70c2bdb23fc13eb6fbc93c9d639fb16cff54b7b76bd3b9c" Mar 21 05:47:03 crc kubenswrapper[4775]: I0321 05:47:03.464515 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-b8t2r/must-gather-znpkx"] Mar 21 05:47:03 crc kubenswrapper[4775]: E0321 05:47:03.465424 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fa06424-b2f2-4965-85d6-52f7824c1663" containerName="extract-content" Mar 21 05:47:03 crc kubenswrapper[4775]: I0321 05:47:03.465445 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fa06424-b2f2-4965-85d6-52f7824c1663" containerName="extract-content" Mar 21 05:47:03 crc kubenswrapper[4775]: E0321 05:47:03.465454 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fa06424-b2f2-4965-85d6-52f7824c1663" containerName="extract-utilities" Mar 21 05:47:03 crc kubenswrapper[4775]: I0321 05:47:03.465463 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fa06424-b2f2-4965-85d6-52f7824c1663" containerName="extract-utilities" Mar 21 05:47:03 crc kubenswrapper[4775]: E0321 05:47:03.465475 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fa06424-b2f2-4965-85d6-52f7824c1663" containerName="registry-server" Mar 21 05:47:03 crc kubenswrapper[4775]: I0321 05:47:03.465484 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fa06424-b2f2-4965-85d6-52f7824c1663" containerName="registry-server" Mar 21 05:47:03 crc kubenswrapper[4775]: I0321 05:47:03.465702 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fa06424-b2f2-4965-85d6-52f7824c1663" containerName="registry-server" Mar 21 05:47:03 crc kubenswrapper[4775]: I0321 05:47:03.467042 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b8t2r/must-gather-znpkx" Mar 21 05:47:03 crc kubenswrapper[4775]: I0321 05:47:03.468496 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-b8t2r"/"default-dockercfg-wsz9m" Mar 21 05:47:03 crc kubenswrapper[4775]: I0321 05:47:03.470952 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-b8t2r"/"openshift-service-ca.crt" Mar 21 05:47:03 crc kubenswrapper[4775]: I0321 05:47:03.471767 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-b8t2r"/"kube-root-ca.crt" Mar 21 05:47:03 crc kubenswrapper[4775]: I0321 05:47:03.491478 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-b8t2r/must-gather-znpkx"] Mar 21 05:47:03 crc kubenswrapper[4775]: I0321 05:47:03.527227 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd8wn\" (UniqueName: \"kubernetes.io/projected/77cb0b91-097f-4e81-888d-57b2d37399bd-kube-api-access-cd8wn\") pod \"must-gather-znpkx\" (UID: \"77cb0b91-097f-4e81-888d-57b2d37399bd\") " pod="openshift-must-gather-b8t2r/must-gather-znpkx" Mar 21 05:47:03 crc kubenswrapper[4775]: I0321 05:47:03.527330 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/77cb0b91-097f-4e81-888d-57b2d37399bd-must-gather-output\") pod \"must-gather-znpkx\" (UID: \"77cb0b91-097f-4e81-888d-57b2d37399bd\") " pod="openshift-must-gather-b8t2r/must-gather-znpkx" Mar 21 05:47:03 crc kubenswrapper[4775]: I0321 05:47:03.629090 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd8wn\" (UniqueName: \"kubernetes.io/projected/77cb0b91-097f-4e81-888d-57b2d37399bd-kube-api-access-cd8wn\") pod \"must-gather-znpkx\" (UID: \"77cb0b91-097f-4e81-888d-57b2d37399bd\") " pod="openshift-must-gather-b8t2r/must-gather-znpkx" Mar 21 05:47:03 crc kubenswrapper[4775]: I0321 05:47:03.629192 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/77cb0b91-097f-4e81-888d-57b2d37399bd-must-gather-output\") pod \"must-gather-znpkx\" (UID: \"77cb0b91-097f-4e81-888d-57b2d37399bd\") " pod="openshift-must-gather-b8t2r/must-gather-znpkx" Mar 21 05:47:03 crc kubenswrapper[4775]: I0321 05:47:03.629674 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/77cb0b91-097f-4e81-888d-57b2d37399bd-must-gather-output\") pod \"must-gather-znpkx\" (UID: \"77cb0b91-097f-4e81-888d-57b2d37399bd\") " pod="openshift-must-gather-b8t2r/must-gather-znpkx" Mar 21 05:47:03 crc kubenswrapper[4775]: I0321 05:47:03.656810 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd8wn\" (UniqueName: \"kubernetes.io/projected/77cb0b91-097f-4e81-888d-57b2d37399bd-kube-api-access-cd8wn\") pod \"must-gather-znpkx\" (UID: \"77cb0b91-097f-4e81-888d-57b2d37399bd\") " pod="openshift-must-gather-b8t2r/must-gather-znpkx" Mar 21 05:47:03 crc kubenswrapper[4775]: I0321 05:47:03.673339 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fa06424-b2f2-4965-85d6-52f7824c1663" path="/var/lib/kubelet/pods/7fa06424-b2f2-4965-85d6-52f7824c1663/volumes" Mar 21 05:47:03 crc kubenswrapper[4775]: I0321 05:47:03.783869 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b8t2r/must-gather-znpkx" Mar 21 05:47:04 crc kubenswrapper[4775]: I0321 05:47:04.356672 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-b8t2r/must-gather-znpkx"] Mar 21 05:47:04 crc kubenswrapper[4775]: I0321 05:47:04.430233 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b8t2r/must-gather-znpkx" event={"ID":"77cb0b91-097f-4e81-888d-57b2d37399bd","Type":"ContainerStarted","Data":"9283ebedf4adbab35d77585b5d52cc34e6dca3a8e4311ff13ca4f606f2d25f69"} Mar 21 05:47:14 crc kubenswrapper[4775]: I0321 05:47:14.519322 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b8t2r/must-gather-znpkx" event={"ID":"77cb0b91-097f-4e81-888d-57b2d37399bd","Type":"ContainerStarted","Data":"76eca6bf5897fa3d0bd946a5ac8c7c93a3f123f745216d5491ccbb010515a8cc"} Mar 21 05:47:14 crc kubenswrapper[4775]: I0321 05:47:14.519951 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b8t2r/must-gather-znpkx" event={"ID":"77cb0b91-097f-4e81-888d-57b2d37399bd","Type":"ContainerStarted","Data":"891104e8ee294ddf739928b04e0a1b0e8f6bf51e7a90dfe50eb44a5a6ed19b18"} Mar 21 05:47:14 crc kubenswrapper[4775]: I0321 05:47:14.542243 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-b8t2r/must-gather-znpkx" podStartSLOduration=2.451127172 podStartE2EDuration="11.542221378s" podCreationTimestamp="2026-03-21 05:47:03 +0000 UTC" firstStartedPulling="2026-03-21 05:47:04.364784907 +0000 UTC m=+3577.341248531" lastFinishedPulling="2026-03-21 05:47:13.455879113 +0000 UTC m=+3586.432342737" observedRunningTime="2026-03-21 05:47:14.534866008 +0000 UTC m=+3587.511329642" watchObservedRunningTime="2026-03-21 05:47:14.542221378 +0000 UTC m=+3587.518685002" Mar 21 05:47:18 crc kubenswrapper[4775]: I0321 05:47:18.553094 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-b8t2r/crc-debug-hwlzx"] Mar 21 05:47:18 crc kubenswrapper[4775]: I0321 05:47:18.556893 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b8t2r/crc-debug-hwlzx" Mar 21 05:47:18 crc kubenswrapper[4775]: I0321 05:47:18.586834 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqnqz\" (UniqueName: \"kubernetes.io/projected/e4c91c96-06cc-4b45-ae11-64b80671e80b-kube-api-access-kqnqz\") pod \"crc-debug-hwlzx\" (UID: \"e4c91c96-06cc-4b45-ae11-64b80671e80b\") " pod="openshift-must-gather-b8t2r/crc-debug-hwlzx" Mar 21 05:47:18 crc kubenswrapper[4775]: I0321 05:47:18.587007 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e4c91c96-06cc-4b45-ae11-64b80671e80b-host\") pod \"crc-debug-hwlzx\" (UID: \"e4c91c96-06cc-4b45-ae11-64b80671e80b\") " pod="openshift-must-gather-b8t2r/crc-debug-hwlzx" Mar 21 05:47:18 crc kubenswrapper[4775]: I0321 05:47:18.689043 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e4c91c96-06cc-4b45-ae11-64b80671e80b-host\") pod \"crc-debug-hwlzx\" (UID: \"e4c91c96-06cc-4b45-ae11-64b80671e80b\") " pod="openshift-must-gather-b8t2r/crc-debug-hwlzx" Mar 21 05:47:18 crc kubenswrapper[4775]: I0321 05:47:18.689296 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqnqz\" (UniqueName: \"kubernetes.io/projected/e4c91c96-06cc-4b45-ae11-64b80671e80b-kube-api-access-kqnqz\") pod \"crc-debug-hwlzx\" (UID: \"e4c91c96-06cc-4b45-ae11-64b80671e80b\") " pod="openshift-must-gather-b8t2r/crc-debug-hwlzx" Mar 21 05:47:18 crc kubenswrapper[4775]: I0321 05:47:18.689652 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e4c91c96-06cc-4b45-ae11-64b80671e80b-host\") pod \"crc-debug-hwlzx\" (UID: \"e4c91c96-06cc-4b45-ae11-64b80671e80b\") " pod="openshift-must-gather-b8t2r/crc-debug-hwlzx" Mar 21 05:47:18 crc kubenswrapper[4775]: I0321 05:47:18.718619 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqnqz\" (UniqueName: \"kubernetes.io/projected/e4c91c96-06cc-4b45-ae11-64b80671e80b-kube-api-access-kqnqz\") pod \"crc-debug-hwlzx\" (UID: \"e4c91c96-06cc-4b45-ae11-64b80671e80b\") " pod="openshift-must-gather-b8t2r/crc-debug-hwlzx" Mar 21 05:47:18 crc kubenswrapper[4775]: I0321 05:47:18.878539 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b8t2r/crc-debug-hwlzx" Mar 21 05:47:18 crc kubenswrapper[4775]: W0321 05:47:18.914826 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4c91c96_06cc_4b45_ae11_64b80671e80b.slice/crio-c614da3ddde7ddcfaca9a6477d42e9beb54e92477909095bbe7803e50891019d WatchSource:0}: Error finding container c614da3ddde7ddcfaca9a6477d42e9beb54e92477909095bbe7803e50891019d: Status 404 returned error can't find the container with id c614da3ddde7ddcfaca9a6477d42e9beb54e92477909095bbe7803e50891019d Mar 21 05:47:19 crc kubenswrapper[4775]: I0321 05:47:19.573542 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b8t2r/crc-debug-hwlzx" event={"ID":"e4c91c96-06cc-4b45-ae11-64b80671e80b","Type":"ContainerStarted","Data":"c614da3ddde7ddcfaca9a6477d42e9beb54e92477909095bbe7803e50891019d"} Mar 21 05:47:35 crc kubenswrapper[4775]: E0321 05:47:35.686486 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296" Mar 21 05:47:35 crc kubenswrapper[4775]: E0321 05:47:35.687254 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:container-00,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296,Command:[chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ; rm -rf \"/var/tmp/sos-osp\" && mkdir -p \"/var/tmp/sos-osp\" && sudo podman rm --force toolbox-osp; sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools && toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=\"/var/tmp/sos-osp\" && if [[ \"$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)\" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf \"/var/tmp/sos-osp/podlogs.tar.xz\" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TMOUT,Value:900,ValueFrom:nil,},EnvVar{Name:HOST,Value:/host,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host,ReadOnly:false,MountPath:/host,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kqnqz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod crc-debug-hwlzx_openshift-must-gather-b8t2r(e4c91c96-06cc-4b45-ae11-64b80671e80b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 05:47:35 crc kubenswrapper[4775]: E0321 05:47:35.688379 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-must-gather-b8t2r/crc-debug-hwlzx" podUID="e4c91c96-06cc-4b45-ae11-64b80671e80b" Mar 21 05:47:35 crc kubenswrapper[4775]: E0321 05:47:35.743756 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296\\\"\"" pod="openshift-must-gather-b8t2r/crc-debug-hwlzx" podUID="e4c91c96-06cc-4b45-ae11-64b80671e80b" Mar 21 05:47:48 crc kubenswrapper[4775]: I0321 05:47:48.859523 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b8t2r/crc-debug-hwlzx" event={"ID":"e4c91c96-06cc-4b45-ae11-64b80671e80b","Type":"ContainerStarted","Data":"59c30d6d373f941e5b2724ff5fbfdf61ac691e4e5c44b0b8bd0b70a7f5049699"} Mar 21 05:47:48 crc kubenswrapper[4775]: I0321 05:47:48.883221 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-b8t2r/crc-debug-hwlzx" podStartSLOduration=1.930702717 podStartE2EDuration="30.88319701s" podCreationTimestamp="2026-03-21 05:47:18 +0000 UTC" firstStartedPulling="2026-03-21 05:47:18.920639969 +0000 UTC m=+3591.897103593" lastFinishedPulling="2026-03-21 05:47:47.873134262 +0000 UTC m=+3620.849597886" observedRunningTime="2026-03-21 05:47:48.881262715 +0000 UTC m=+3621.857726349" watchObservedRunningTime="2026-03-21 05:47:48.88319701 +0000 UTC m=+3621.859660634" Mar 21 05:48:00 crc kubenswrapper[4775]: I0321 05:48:00.149011 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567868-8z86d"] Mar 21 05:48:00 crc kubenswrapper[4775]: I0321 05:48:00.151348 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567868-8z86d" Mar 21 05:48:00 crc kubenswrapper[4775]: I0321 05:48:00.153676 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:48:00 crc kubenswrapper[4775]: I0321 05:48:00.154984 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:48:00 crc kubenswrapper[4775]: I0321 05:48:00.156134 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:48:00 crc kubenswrapper[4775]: I0321 05:48:00.163739 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567868-8z86d"] Mar 21 05:48:00 crc kubenswrapper[4775]: I0321 05:48:00.212786 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dsp9\" (UniqueName: \"kubernetes.io/projected/bdeb3ead-3f1f-491f-9a15-f550065d18fb-kube-api-access-5dsp9\") pod \"auto-csr-approver-29567868-8z86d\" (UID: \"bdeb3ead-3f1f-491f-9a15-f550065d18fb\") " pod="openshift-infra/auto-csr-approver-29567868-8z86d" Mar 21 05:48:00 crc kubenswrapper[4775]: I0321 05:48:00.315762 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dsp9\" (UniqueName: \"kubernetes.io/projected/bdeb3ead-3f1f-491f-9a15-f550065d18fb-kube-api-access-5dsp9\") pod \"auto-csr-approver-29567868-8z86d\" (UID: \"bdeb3ead-3f1f-491f-9a15-f550065d18fb\") " pod="openshift-infra/auto-csr-approver-29567868-8z86d" Mar 21 05:48:00 crc kubenswrapper[4775]: I0321 05:48:00.342156 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dsp9\" (UniqueName: \"kubernetes.io/projected/bdeb3ead-3f1f-491f-9a15-f550065d18fb-kube-api-access-5dsp9\") pod \"auto-csr-approver-29567868-8z86d\" (UID: \"bdeb3ead-3f1f-491f-9a15-f550065d18fb\") " pod="openshift-infra/auto-csr-approver-29567868-8z86d" Mar 21 05:48:00 crc kubenswrapper[4775]: I0321 05:48:00.864613 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567868-8z86d" Mar 21 05:48:01 crc kubenswrapper[4775]: I0321 05:48:01.352563 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567868-8z86d"] Mar 21 05:48:02 crc kubenswrapper[4775]: I0321 05:48:02.007330 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567868-8z86d" event={"ID":"bdeb3ead-3f1f-491f-9a15-f550065d18fb","Type":"ContainerStarted","Data":"7d50155bebefa4c1f1d1c9ef73ca60fb016430db6c5e51096e9341267622df9b"} Mar 21 05:48:02 crc kubenswrapper[4775]: I0321 05:48:02.481933 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:48:02 crc kubenswrapper[4775]: I0321 05:48:02.482009 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:48:04 crc kubenswrapper[4775]: I0321 05:48:04.030138 4775 generic.go:334] "Generic (PLEG): container finished" podID="bdeb3ead-3f1f-491f-9a15-f550065d18fb" containerID="6995f01e339d899d20a0310c4264870e5b4687e152b20b7bc6f33c17d3fccaea" exitCode=0 Mar 21 05:48:04 crc kubenswrapper[4775]: I0321 05:48:04.030245 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567868-8z86d" event={"ID":"bdeb3ead-3f1f-491f-9a15-f550065d18fb","Type":"ContainerDied","Data":"6995f01e339d899d20a0310c4264870e5b4687e152b20b7bc6f33c17d3fccaea"} Mar 21 05:48:05 crc kubenswrapper[4775]: I0321 05:48:05.410038 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567868-8z86d" Mar 21 05:48:05 crc kubenswrapper[4775]: I0321 05:48:05.521543 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dsp9\" (UniqueName: \"kubernetes.io/projected/bdeb3ead-3f1f-491f-9a15-f550065d18fb-kube-api-access-5dsp9\") pod \"bdeb3ead-3f1f-491f-9a15-f550065d18fb\" (UID: \"bdeb3ead-3f1f-491f-9a15-f550065d18fb\") " Mar 21 05:48:05 crc kubenswrapper[4775]: I0321 05:48:05.534343 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdeb3ead-3f1f-491f-9a15-f550065d18fb-kube-api-access-5dsp9" (OuterVolumeSpecName: "kube-api-access-5dsp9") pod "bdeb3ead-3f1f-491f-9a15-f550065d18fb" (UID: "bdeb3ead-3f1f-491f-9a15-f550065d18fb"). InnerVolumeSpecName "kube-api-access-5dsp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:48:05 crc kubenswrapper[4775]: I0321 05:48:05.624028 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dsp9\" (UniqueName: \"kubernetes.io/projected/bdeb3ead-3f1f-491f-9a15-f550065d18fb-kube-api-access-5dsp9\") on node \"crc\" DevicePath \"\"" Mar 21 05:48:06 crc kubenswrapper[4775]: I0321 05:48:06.052178 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567868-8z86d" event={"ID":"bdeb3ead-3f1f-491f-9a15-f550065d18fb","Type":"ContainerDied","Data":"7d50155bebefa4c1f1d1c9ef73ca60fb016430db6c5e51096e9341267622df9b"} Mar 21 05:48:06 crc kubenswrapper[4775]: I0321 05:48:06.052218 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d50155bebefa4c1f1d1c9ef73ca60fb016430db6c5e51096e9341267622df9b" Mar 21 05:48:06 crc kubenswrapper[4775]: I0321 05:48:06.052255 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567868-8z86d" Mar 21 05:48:06 crc kubenswrapper[4775]: I0321 05:48:06.494885 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567862-bbnhp"] Mar 21 05:48:06 crc kubenswrapper[4775]: I0321 05:48:06.504973 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567862-bbnhp"] Mar 21 05:48:07 crc kubenswrapper[4775]: I0321 05:48:07.670827 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89f5316a-d6df-4c37-b878-4f3ef82d837b" path="/var/lib/kubelet/pods/89f5316a-d6df-4c37-b878-4f3ef82d837b/volumes" Mar 21 05:48:26 crc kubenswrapper[4775]: I0321 05:48:26.286230 4775 generic.go:334] "Generic (PLEG): container finished" podID="e4c91c96-06cc-4b45-ae11-64b80671e80b" containerID="59c30d6d373f941e5b2724ff5fbfdf61ac691e4e5c44b0b8bd0b70a7f5049699" exitCode=0 Mar 21 05:48:26 crc kubenswrapper[4775]: I0321 05:48:26.286318 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b8t2r/crc-debug-hwlzx" event={"ID":"e4c91c96-06cc-4b45-ae11-64b80671e80b","Type":"ContainerDied","Data":"59c30d6d373f941e5b2724ff5fbfdf61ac691e4e5c44b0b8bd0b70a7f5049699"} Mar 21 05:48:27 crc kubenswrapper[4775]: I0321 05:48:27.411589 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b8t2r/crc-debug-hwlzx" Mar 21 05:48:27 crc kubenswrapper[4775]: I0321 05:48:27.451394 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-b8t2r/crc-debug-hwlzx"] Mar 21 05:48:27 crc kubenswrapper[4775]: I0321 05:48:27.459727 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-b8t2r/crc-debug-hwlzx"] Mar 21 05:48:27 crc kubenswrapper[4775]: I0321 05:48:27.486490 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e4c91c96-06cc-4b45-ae11-64b80671e80b-host\") pod \"e4c91c96-06cc-4b45-ae11-64b80671e80b\" (UID: \"e4c91c96-06cc-4b45-ae11-64b80671e80b\") " Mar 21 05:48:27 crc kubenswrapper[4775]: I0321 05:48:27.486614 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e4c91c96-06cc-4b45-ae11-64b80671e80b-host" (OuterVolumeSpecName: "host") pod "e4c91c96-06cc-4b45-ae11-64b80671e80b" (UID: "e4c91c96-06cc-4b45-ae11-64b80671e80b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:48:27 crc kubenswrapper[4775]: I0321 05:48:27.486640 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqnqz\" (UniqueName: \"kubernetes.io/projected/e4c91c96-06cc-4b45-ae11-64b80671e80b-kube-api-access-kqnqz\") pod \"e4c91c96-06cc-4b45-ae11-64b80671e80b\" (UID: \"e4c91c96-06cc-4b45-ae11-64b80671e80b\") " Mar 21 05:48:27 crc kubenswrapper[4775]: I0321 05:48:27.487013 4775 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e4c91c96-06cc-4b45-ae11-64b80671e80b-host\") on node \"crc\" DevicePath \"\"" Mar 21 05:48:27 crc kubenswrapper[4775]: I0321 05:48:27.491985 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4c91c96-06cc-4b45-ae11-64b80671e80b-kube-api-access-kqnqz" (OuterVolumeSpecName: "kube-api-access-kqnqz") pod "e4c91c96-06cc-4b45-ae11-64b80671e80b" (UID: "e4c91c96-06cc-4b45-ae11-64b80671e80b"). InnerVolumeSpecName "kube-api-access-kqnqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:48:27 crc kubenswrapper[4775]: I0321 05:48:27.588772 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqnqz\" (UniqueName: \"kubernetes.io/projected/e4c91c96-06cc-4b45-ae11-64b80671e80b-kube-api-access-kqnqz\") on node \"crc\" DevicePath \"\"" Mar 21 05:48:27 crc kubenswrapper[4775]: I0321 05:48:27.678729 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4c91c96-06cc-4b45-ae11-64b80671e80b" path="/var/lib/kubelet/pods/e4c91c96-06cc-4b45-ae11-64b80671e80b/volumes" Mar 21 05:48:28 crc kubenswrapper[4775]: I0321 05:48:28.305232 4775 scope.go:117] "RemoveContainer" containerID="59c30d6d373f941e5b2724ff5fbfdf61ac691e4e5c44b0b8bd0b70a7f5049699" Mar 21 05:48:28 crc kubenswrapper[4775]: I0321 05:48:28.305284 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b8t2r/crc-debug-hwlzx" Mar 21 05:48:28 crc kubenswrapper[4775]: I0321 05:48:28.614787 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-b8t2r/crc-debug-895n9"] Mar 21 05:48:28 crc kubenswrapper[4775]: E0321 05:48:28.615283 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4c91c96-06cc-4b45-ae11-64b80671e80b" containerName="container-00" Mar 21 05:48:28 crc kubenswrapper[4775]: I0321 05:48:28.615299 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4c91c96-06cc-4b45-ae11-64b80671e80b" containerName="container-00" Mar 21 05:48:28 crc kubenswrapper[4775]: E0321 05:48:28.615326 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdeb3ead-3f1f-491f-9a15-f550065d18fb" containerName="oc" Mar 21 05:48:28 crc kubenswrapper[4775]: I0321 05:48:28.615337 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdeb3ead-3f1f-491f-9a15-f550065d18fb" containerName="oc" Mar 21 05:48:28 crc kubenswrapper[4775]: I0321 05:48:28.615634 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdeb3ead-3f1f-491f-9a15-f550065d18fb" containerName="oc" Mar 21 05:48:28 crc kubenswrapper[4775]: I0321 05:48:28.615666 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4c91c96-06cc-4b45-ae11-64b80671e80b" containerName="container-00" Mar 21 05:48:28 crc kubenswrapper[4775]: I0321 05:48:28.617108 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b8t2r/crc-debug-895n9" Mar 21 05:48:28 crc kubenswrapper[4775]: I0321 05:48:28.712669 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxz5n\" (UniqueName: \"kubernetes.io/projected/4fb0eb42-ffda-49af-a49c-d9ecdb7af97f-kube-api-access-sxz5n\") pod \"crc-debug-895n9\" (UID: \"4fb0eb42-ffda-49af-a49c-d9ecdb7af97f\") " pod="openshift-must-gather-b8t2r/crc-debug-895n9" Mar 21 05:48:28 crc kubenswrapper[4775]: I0321 05:48:28.712873 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4fb0eb42-ffda-49af-a49c-d9ecdb7af97f-host\") pod \"crc-debug-895n9\" (UID: \"4fb0eb42-ffda-49af-a49c-d9ecdb7af97f\") " pod="openshift-must-gather-b8t2r/crc-debug-895n9" Mar 21 05:48:28 crc kubenswrapper[4775]: I0321 05:48:28.815245 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxz5n\" (UniqueName: \"kubernetes.io/projected/4fb0eb42-ffda-49af-a49c-d9ecdb7af97f-kube-api-access-sxz5n\") pod \"crc-debug-895n9\" (UID: \"4fb0eb42-ffda-49af-a49c-d9ecdb7af97f\") " pod="openshift-must-gather-b8t2r/crc-debug-895n9" Mar 21 05:48:28 crc kubenswrapper[4775]: I0321 05:48:28.815368 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4fb0eb42-ffda-49af-a49c-d9ecdb7af97f-host\") pod \"crc-debug-895n9\" (UID: \"4fb0eb42-ffda-49af-a49c-d9ecdb7af97f\") " pod="openshift-must-gather-b8t2r/crc-debug-895n9" Mar 21 05:48:28 crc kubenswrapper[4775]: I0321 05:48:28.815707 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4fb0eb42-ffda-49af-a49c-d9ecdb7af97f-host\") pod \"crc-debug-895n9\" (UID: \"4fb0eb42-ffda-49af-a49c-d9ecdb7af97f\") " pod="openshift-must-gather-b8t2r/crc-debug-895n9" Mar 21 05:48:28 crc kubenswrapper[4775]: I0321 05:48:28.839274 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxz5n\" (UniqueName: \"kubernetes.io/projected/4fb0eb42-ffda-49af-a49c-d9ecdb7af97f-kube-api-access-sxz5n\") pod \"crc-debug-895n9\" (UID: \"4fb0eb42-ffda-49af-a49c-d9ecdb7af97f\") " pod="openshift-must-gather-b8t2r/crc-debug-895n9" Mar 21 05:48:28 crc kubenswrapper[4775]: I0321 05:48:28.937267 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b8t2r/crc-debug-895n9" Mar 21 05:48:28 crc kubenswrapper[4775]: W0321 05:48:28.966817 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4fb0eb42_ffda_49af_a49c_d9ecdb7af97f.slice/crio-28871b09be768f444462e0b438b5c3800defbdffe8f0e221f6f09be59d8d2535 WatchSource:0}: Error finding container 28871b09be768f444462e0b438b5c3800defbdffe8f0e221f6f09be59d8d2535: Status 404 returned error can't find the container with id 28871b09be768f444462e0b438b5c3800defbdffe8f0e221f6f09be59d8d2535 Mar 21 05:48:29 crc kubenswrapper[4775]: I0321 05:48:29.317661 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b8t2r/crc-debug-895n9" event={"ID":"4fb0eb42-ffda-49af-a49c-d9ecdb7af97f","Type":"ContainerStarted","Data":"f3085622e93162df966234d5c6f8955f876617fcb1bf4d233a2caaceadb9d60f"} Mar 21 05:48:29 crc kubenswrapper[4775]: I0321 05:48:29.317763 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b8t2r/crc-debug-895n9" event={"ID":"4fb0eb42-ffda-49af-a49c-d9ecdb7af97f","Type":"ContainerStarted","Data":"28871b09be768f444462e0b438b5c3800defbdffe8f0e221f6f09be59d8d2535"} Mar 21 05:48:29 crc kubenswrapper[4775]: I0321 05:48:29.336072 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-b8t2r/crc-debug-895n9" podStartSLOduration=1.336038829 podStartE2EDuration="1.336038829s" podCreationTimestamp="2026-03-21 05:48:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:48:29.331765388 +0000 UTC m=+3662.308229032" watchObservedRunningTime="2026-03-21 05:48:29.336038829 +0000 UTC m=+3662.312502493" Mar 21 05:48:30 crc kubenswrapper[4775]: I0321 05:48:30.327977 4775 generic.go:334] "Generic (PLEG): container finished" podID="4fb0eb42-ffda-49af-a49c-d9ecdb7af97f" containerID="f3085622e93162df966234d5c6f8955f876617fcb1bf4d233a2caaceadb9d60f" exitCode=0 Mar 21 05:48:30 crc kubenswrapper[4775]: I0321 05:48:30.328051 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b8t2r/crc-debug-895n9" event={"ID":"4fb0eb42-ffda-49af-a49c-d9ecdb7af97f","Type":"ContainerDied","Data":"f3085622e93162df966234d5c6f8955f876617fcb1bf4d233a2caaceadb9d60f"} Mar 21 05:48:31 crc kubenswrapper[4775]: I0321 05:48:31.440649 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b8t2r/crc-debug-895n9" Mar 21 05:48:31 crc kubenswrapper[4775]: I0321 05:48:31.470579 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-b8t2r/crc-debug-895n9"] Mar 21 05:48:31 crc kubenswrapper[4775]: I0321 05:48:31.478714 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-b8t2r/crc-debug-895n9"] Mar 21 05:48:31 crc kubenswrapper[4775]: I0321 05:48:31.585571 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxz5n\" (UniqueName: \"kubernetes.io/projected/4fb0eb42-ffda-49af-a49c-d9ecdb7af97f-kube-api-access-sxz5n\") pod \"4fb0eb42-ffda-49af-a49c-d9ecdb7af97f\" (UID: \"4fb0eb42-ffda-49af-a49c-d9ecdb7af97f\") " Mar 21 05:48:31 crc kubenswrapper[4775]: I0321 05:48:31.585743 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4fb0eb42-ffda-49af-a49c-d9ecdb7af97f-host\") pod \"4fb0eb42-ffda-49af-a49c-d9ecdb7af97f\" (UID: \"4fb0eb42-ffda-49af-a49c-d9ecdb7af97f\") " Mar 21 05:48:31 crc kubenswrapper[4775]: I0321 05:48:31.585894 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4fb0eb42-ffda-49af-a49c-d9ecdb7af97f-host" (OuterVolumeSpecName: "host") pod "4fb0eb42-ffda-49af-a49c-d9ecdb7af97f" (UID: "4fb0eb42-ffda-49af-a49c-d9ecdb7af97f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:48:31 crc kubenswrapper[4775]: I0321 05:48:31.586438 4775 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4fb0eb42-ffda-49af-a49c-d9ecdb7af97f-host\") on node \"crc\" DevicePath \"\"" Mar 21 05:48:31 crc kubenswrapper[4775]: I0321 05:48:31.594481 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fb0eb42-ffda-49af-a49c-d9ecdb7af97f-kube-api-access-sxz5n" (OuterVolumeSpecName: "kube-api-access-sxz5n") pod "4fb0eb42-ffda-49af-a49c-d9ecdb7af97f" (UID: "4fb0eb42-ffda-49af-a49c-d9ecdb7af97f"). InnerVolumeSpecName "kube-api-access-sxz5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:48:31 crc kubenswrapper[4775]: I0321 05:48:31.675024 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fb0eb42-ffda-49af-a49c-d9ecdb7af97f" path="/var/lib/kubelet/pods/4fb0eb42-ffda-49af-a49c-d9ecdb7af97f/volumes" Mar 21 05:48:31 crc kubenswrapper[4775]: I0321 05:48:31.688294 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxz5n\" (UniqueName: \"kubernetes.io/projected/4fb0eb42-ffda-49af-a49c-d9ecdb7af97f-kube-api-access-sxz5n\") on node \"crc\" DevicePath \"\"" Mar 21 05:48:32 crc kubenswrapper[4775]: I0321 05:48:32.347700 4775 scope.go:117] "RemoveContainer" containerID="f3085622e93162df966234d5c6f8955f876617fcb1bf4d233a2caaceadb9d60f" Mar 21 05:48:32 crc kubenswrapper[4775]: I0321 05:48:32.347755 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b8t2r/crc-debug-895n9" Mar 21 05:48:32 crc kubenswrapper[4775]: I0321 05:48:32.481952 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:48:32 crc kubenswrapper[4775]: I0321 05:48:32.482667 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:48:32 crc kubenswrapper[4775]: I0321 05:48:32.642063 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-b8t2r/crc-debug-k6ct4"] Mar 21 05:48:32 crc kubenswrapper[4775]: E0321 05:48:32.642811 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fb0eb42-ffda-49af-a49c-d9ecdb7af97f" containerName="container-00" Mar 21 05:48:32 crc kubenswrapper[4775]: I0321 05:48:32.642829 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fb0eb42-ffda-49af-a49c-d9ecdb7af97f" containerName="container-00" Mar 21 05:48:32 crc kubenswrapper[4775]: I0321 05:48:32.643014 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fb0eb42-ffda-49af-a49c-d9ecdb7af97f" containerName="container-00" Mar 21 05:48:32 crc kubenswrapper[4775]: I0321 05:48:32.643728 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b8t2r/crc-debug-k6ct4" Mar 21 05:48:32 crc kubenswrapper[4775]: I0321 05:48:32.810506 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c04654f3-0f1c-4883-8385-4149f0076d44-host\") pod \"crc-debug-k6ct4\" (UID: \"c04654f3-0f1c-4883-8385-4149f0076d44\") " pod="openshift-must-gather-b8t2r/crc-debug-k6ct4" Mar 21 05:48:32 crc kubenswrapper[4775]: I0321 05:48:32.811031 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9mn2\" (UniqueName: \"kubernetes.io/projected/c04654f3-0f1c-4883-8385-4149f0076d44-kube-api-access-m9mn2\") pod \"crc-debug-k6ct4\" (UID: \"c04654f3-0f1c-4883-8385-4149f0076d44\") " pod="openshift-must-gather-b8t2r/crc-debug-k6ct4" Mar 21 05:48:32 crc kubenswrapper[4775]: I0321 05:48:32.913403 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c04654f3-0f1c-4883-8385-4149f0076d44-host\") pod \"crc-debug-k6ct4\" (UID: \"c04654f3-0f1c-4883-8385-4149f0076d44\") " pod="openshift-must-gather-b8t2r/crc-debug-k6ct4" Mar 21 05:48:32 crc kubenswrapper[4775]: I0321 05:48:32.913451 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9mn2\" (UniqueName: \"kubernetes.io/projected/c04654f3-0f1c-4883-8385-4149f0076d44-kube-api-access-m9mn2\") pod \"crc-debug-k6ct4\" (UID: \"c04654f3-0f1c-4883-8385-4149f0076d44\") " pod="openshift-must-gather-b8t2r/crc-debug-k6ct4" Mar 21 05:48:32 crc kubenswrapper[4775]: I0321 05:48:32.913613 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c04654f3-0f1c-4883-8385-4149f0076d44-host\") pod \"crc-debug-k6ct4\" (UID: \"c04654f3-0f1c-4883-8385-4149f0076d44\") " pod="openshift-must-gather-b8t2r/crc-debug-k6ct4" Mar 21 05:48:32 crc kubenswrapper[4775]: I0321 05:48:32.934035 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9mn2\" (UniqueName: \"kubernetes.io/projected/c04654f3-0f1c-4883-8385-4149f0076d44-kube-api-access-m9mn2\") pod \"crc-debug-k6ct4\" (UID: \"c04654f3-0f1c-4883-8385-4149f0076d44\") " pod="openshift-must-gather-b8t2r/crc-debug-k6ct4" Mar 21 05:48:32 crc kubenswrapper[4775]: I0321 05:48:32.961342 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b8t2r/crc-debug-k6ct4" Mar 21 05:48:32 crc kubenswrapper[4775]: W0321 05:48:32.997022 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc04654f3_0f1c_4883_8385_4149f0076d44.slice/crio-5c91c61fdedd7ac3b55f157313b39d5da2758e7741b49a653fbe9b5afd78473b WatchSource:0}: Error finding container 5c91c61fdedd7ac3b55f157313b39d5da2758e7741b49a653fbe9b5afd78473b: Status 404 returned error can't find the container with id 5c91c61fdedd7ac3b55f157313b39d5da2758e7741b49a653fbe9b5afd78473b Mar 21 05:48:33 crc kubenswrapper[4775]: I0321 05:48:33.360078 4775 generic.go:334] "Generic (PLEG): container finished" podID="c04654f3-0f1c-4883-8385-4149f0076d44" containerID="7531509e6b8b22386d2d189b980c4f48fe41789688a7cacac28bf362ed5c6b17" exitCode=0 Mar 21 05:48:33 crc kubenswrapper[4775]: I0321 05:48:33.360167 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b8t2r/crc-debug-k6ct4" event={"ID":"c04654f3-0f1c-4883-8385-4149f0076d44","Type":"ContainerDied","Data":"7531509e6b8b22386d2d189b980c4f48fe41789688a7cacac28bf362ed5c6b17"} Mar 21 05:48:33 crc kubenswrapper[4775]: I0321 05:48:33.360347 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b8t2r/crc-debug-k6ct4" event={"ID":"c04654f3-0f1c-4883-8385-4149f0076d44","Type":"ContainerStarted","Data":"5c91c61fdedd7ac3b55f157313b39d5da2758e7741b49a653fbe9b5afd78473b"} Mar 21 05:48:33 crc kubenswrapper[4775]: I0321 05:48:33.403088 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-b8t2r/crc-debug-k6ct4"] Mar 21 05:48:33 crc kubenswrapper[4775]: I0321 05:48:33.413500 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-b8t2r/crc-debug-k6ct4"] Mar 21 05:48:34 crc kubenswrapper[4775]: I0321 05:48:34.474095 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b8t2r/crc-debug-k6ct4" Mar 21 05:48:34 crc kubenswrapper[4775]: I0321 05:48:34.647272 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9mn2\" (UniqueName: \"kubernetes.io/projected/c04654f3-0f1c-4883-8385-4149f0076d44-kube-api-access-m9mn2\") pod \"c04654f3-0f1c-4883-8385-4149f0076d44\" (UID: \"c04654f3-0f1c-4883-8385-4149f0076d44\") " Mar 21 05:48:34 crc kubenswrapper[4775]: I0321 05:48:34.647422 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c04654f3-0f1c-4883-8385-4149f0076d44-host\") pod \"c04654f3-0f1c-4883-8385-4149f0076d44\" (UID: \"c04654f3-0f1c-4883-8385-4149f0076d44\") " Mar 21 05:48:34 crc kubenswrapper[4775]: I0321 05:48:34.647853 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c04654f3-0f1c-4883-8385-4149f0076d44-host" (OuterVolumeSpecName: "host") pod "c04654f3-0f1c-4883-8385-4149f0076d44" (UID: "c04654f3-0f1c-4883-8385-4149f0076d44"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:48:34 crc kubenswrapper[4775]: I0321 05:48:34.648177 4775 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c04654f3-0f1c-4883-8385-4149f0076d44-host\") on node \"crc\" DevicePath \"\"" Mar 21 05:48:34 crc kubenswrapper[4775]: I0321 05:48:34.653329 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c04654f3-0f1c-4883-8385-4149f0076d44-kube-api-access-m9mn2" (OuterVolumeSpecName: "kube-api-access-m9mn2") pod "c04654f3-0f1c-4883-8385-4149f0076d44" (UID: "c04654f3-0f1c-4883-8385-4149f0076d44"). InnerVolumeSpecName "kube-api-access-m9mn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:48:34 crc kubenswrapper[4775]: I0321 05:48:34.750411 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9mn2\" (UniqueName: \"kubernetes.io/projected/c04654f3-0f1c-4883-8385-4149f0076d44-kube-api-access-m9mn2\") on node \"crc\" DevicePath \"\"" Mar 21 05:48:35 crc kubenswrapper[4775]: I0321 05:48:35.897379 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c04654f3-0f1c-4883-8385-4149f0076d44" path="/var/lib/kubelet/pods/c04654f3-0f1c-4883-8385-4149f0076d44/volumes" Mar 21 05:48:35 crc kubenswrapper[4775]: I0321 05:48:35.918609 4775 scope.go:117] "RemoveContainer" containerID="7531509e6b8b22386d2d189b980c4f48fe41789688a7cacac28bf362ed5c6b17" Mar 21 05:48:35 crc kubenswrapper[4775]: I0321 05:48:35.918955 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b8t2r/crc-debug-k6ct4" Mar 21 05:48:48 crc kubenswrapper[4775]: I0321 05:48:48.957581 4775 scope.go:117] "RemoveContainer" containerID="5c92c80d8920da595f41beeabde7d0200200f2dfb0bec0c216c4b59513f9ce40" Mar 21 05:48:50 crc kubenswrapper[4775]: I0321 05:48:50.437185 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-77fd86567d-mf2wb_1727040d-36f5-431c-b8f1-84e206146dcf/barbican-api/0.log" Mar 21 05:48:50 crc kubenswrapper[4775]: I0321 05:48:50.572397 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-77fd86567d-mf2wb_1727040d-36f5-431c-b8f1-84e206146dcf/barbican-api-log/0.log" Mar 21 05:48:50 crc kubenswrapper[4775]: I0321 05:48:50.625859 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-75f5b547c8-mgjw5_177228f6-7f69-49c2-9942-ea0a98b56d13/barbican-keystone-listener/0.log" Mar 21 05:48:50 crc kubenswrapper[4775]: I0321 05:48:50.701598 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-75f5b547c8-mgjw5_177228f6-7f69-49c2-9942-ea0a98b56d13/barbican-keystone-listener-log/0.log" Mar 21 05:48:50 crc kubenswrapper[4775]: I0321 05:48:50.916190 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6dd9f89d55-fdf8c_d8a7c2e5-3643-4675-9888-3c310e4f9ad4/barbican-worker/0.log" Mar 21 05:48:50 crc kubenswrapper[4775]: I0321 05:48:50.924071 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6dd9f89d55-fdf8c_d8a7c2e5-3643-4675-9888-3c310e4f9ad4/barbican-worker-log/0.log" Mar 21 05:48:51 crc kubenswrapper[4775]: I0321 05:48:51.242214 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m_203df932-0574-4098-b897-ba50813f2ec1/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:48:51 crc kubenswrapper[4775]: I0321 05:48:51.354015 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_15d97495-428d-47e0-a115-99c7fd08850a/ceilometer-central-agent/0.log" Mar 21 05:48:51 crc kubenswrapper[4775]: I0321 05:48:51.392935 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_15d97495-428d-47e0-a115-99c7fd08850a/ceilometer-notification-agent/0.log" Mar 21 05:48:51 crc kubenswrapper[4775]: I0321 05:48:51.449906 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_15d97495-428d-47e0-a115-99c7fd08850a/proxy-httpd/0.log" Mar 21 05:48:51 crc kubenswrapper[4775]: I0321 05:48:51.485141 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_15d97495-428d-47e0-a115-99c7fd08850a/sg-core/0.log" Mar 21 05:48:51 crc kubenswrapper[4775]: I0321 05:48:51.632277 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c16d835a-1ec2-473d-b2d8-c8e7c978e140/cinder-api/0.log" Mar 21 05:48:51 crc kubenswrapper[4775]: I0321 05:48:51.689903 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c16d835a-1ec2-473d-b2d8-c8e7c978e140/cinder-api-log/0.log" Mar 21 05:48:51 crc kubenswrapper[4775]: I0321 05:48:51.815198 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d02bb319-292e-449c-8e5b-42c6859f1529/cinder-scheduler/0.log" Mar 21 05:48:51 crc kubenswrapper[4775]: I0321 05:48:51.839665 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d02bb319-292e-449c-8e5b-42c6859f1529/probe/0.log" Mar 21 05:48:52 crc kubenswrapper[4775]: I0321 05:48:52.007295 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7_5efe4255-484c-47d7-800a-4d0dbc5cecd9/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:48:52 crc kubenswrapper[4775]: I0321 05:48:52.196594 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn_8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:48:52 crc kubenswrapper[4775]: I0321 05:48:52.258038 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-mmgp9_4918607c-6074-4fb3-a0a0-8def479058a0/init/0.log" Mar 21 05:48:52 crc kubenswrapper[4775]: I0321 05:48:52.493684 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-mmgp9_4918607c-6074-4fb3-a0a0-8def479058a0/init/0.log" Mar 21 05:48:52 crc kubenswrapper[4775]: I0321 05:48:52.558744 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-mmgp9_4918607c-6074-4fb3-a0a0-8def479058a0/dnsmasq-dns/0.log" Mar 21 05:48:52 crc kubenswrapper[4775]: I0321 05:48:52.564350 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-h87nr_28040d61-c9ea-4a55-b113-db871dff679c/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:48:52 crc kubenswrapper[4775]: I0321 05:48:52.756675 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_01638b90-5e17-43b3-a3b5-90726b26e243/glance-httpd/0.log" Mar 21 05:48:52 crc kubenswrapper[4775]: I0321 05:48:52.795261 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_01638b90-5e17-43b3-a3b5-90726b26e243/glance-log/0.log" Mar 21 05:48:52 crc kubenswrapper[4775]: I0321 05:48:52.922510 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_e9e72c4b-a3fb-41eb-974a-74d24d6cdac9/glance-httpd/0.log" Mar 21 05:48:52 crc kubenswrapper[4775]: I0321 05:48:52.956443 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_e9e72c4b-a3fb-41eb-974a-74d24d6cdac9/glance-log/0.log" Mar 21 05:48:53 crc kubenswrapper[4775]: I0321 05:48:53.184975 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6496ddbdd4-v5mc5_fc6e433f-9e70-4b09-9780-403634bbe0dc/horizon/0.log" Mar 21 05:48:53 crc kubenswrapper[4775]: I0321 05:48:53.353740 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f_8390751b-3911-4a24-a1a2-c3d1d10da875/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:48:53 crc kubenswrapper[4775]: I0321 05:48:53.494854 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6496ddbdd4-v5mc5_fc6e433f-9e70-4b09-9780-403634bbe0dc/horizon-log/0.log" Mar 21 05:48:53 crc kubenswrapper[4775]: I0321 05:48:53.735696 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-kfrz2_dbd64e65-be8d-42e7-a686-d5454932156d/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:48:53 crc kubenswrapper[4775]: I0321 05:48:53.881319 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_7ee3b7a0-9eb3-4702-8fb7-3286df60b21b/kube-state-metrics/0.log" Mar 21 05:48:53 crc kubenswrapper[4775]: I0321 05:48:53.913430 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-66c879cfdd-smnxp_9f965feb-5d82-4176-a14d-08a84c4ae794/keystone-api/0.log" Mar 21 05:48:54 crc kubenswrapper[4775]: I0321 05:48:54.525858 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7d65998c7c-prp5b_3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef/neutron-api/0.log" Mar 21 05:48:54 crc kubenswrapper[4775]: I0321 05:48:54.647762 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7d65998c7c-prp5b_3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef/neutron-httpd/0.log" Mar 21 05:48:54 crc kubenswrapper[4775]: I0321 05:48:54.734058 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz_a71dcc90-c70a-4ff8-bf4a-42f1a2415827/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:48:55 crc kubenswrapper[4775]: I0321 05:48:55.012748 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9_29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:48:55 crc kubenswrapper[4775]: I0321 05:48:55.480501 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_789a0bb8-b131-4144-9400-7c32a604d6d5/nova-cell0-conductor-conductor/0.log" Mar 21 05:48:55 crc kubenswrapper[4775]: I0321 05:48:55.542453 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_886c404c-ceec-48e7-90da-96d6aa201152/nova-api-log/0.log" Mar 21 05:48:55 crc kubenswrapper[4775]: I0321 05:48:55.762563 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_886c404c-ceec-48e7-90da-96d6aa201152/nova-api-api/0.log" Mar 21 05:48:55 crc kubenswrapper[4775]: I0321 05:48:55.817790 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_b29323a7-476f-4a13-8085-4b2158a68850/nova-cell1-conductor-conductor/0.log" Mar 21 05:48:55 crc kubenswrapper[4775]: I0321 05:48:55.902220 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_3b7ea443-e30d-41d1-9f42-0bef9d7bd012/nova-cell1-novncproxy-novncproxy/0.log" Mar 21 05:48:56 crc kubenswrapper[4775]: I0321 05:48:56.369532 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_208cfa71-8242-4958-b9db-21fc180a6697/nova-metadata-log/0.log" Mar 21 05:48:56 crc kubenswrapper[4775]: I0321 05:48:56.766795 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-t85b7_50003f97-774c-4321-9ddf-6ac67546b19f/nova-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:48:56 crc kubenswrapper[4775]: I0321 05:48:56.777469 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_208cfa71-8242-4958-b9db-21fc180a6697/nova-metadata-metadata/0.log" Mar 21 05:48:56 crc kubenswrapper[4775]: I0321 05:48:56.801159 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_bab55731-40da-4831-a8b5-f9c413452367/nova-scheduler-scheduler/0.log" Mar 21 05:48:56 crc kubenswrapper[4775]: I0321 05:48:56.918560 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bade9789-f227-44ab-b7fa-2173445cd381/mysql-bootstrap/0.log" Mar 21 05:48:57 crc kubenswrapper[4775]: I0321 05:48:57.111405 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bade9789-f227-44ab-b7fa-2173445cd381/mysql-bootstrap/0.log" Mar 21 05:48:57 crc kubenswrapper[4775]: I0321 05:48:57.212828 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_76d205b7-bc2e-4dad-b513-457ff20d67e1/mysql-bootstrap/0.log" Mar 21 05:48:57 crc kubenswrapper[4775]: I0321 05:48:57.231599 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bade9789-f227-44ab-b7fa-2173445cd381/galera/0.log" Mar 21 05:48:57 crc kubenswrapper[4775]: I0321 05:48:57.434940 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_76d205b7-bc2e-4dad-b513-457ff20d67e1/mysql-bootstrap/0.log" Mar 21 05:48:57 crc kubenswrapper[4775]: I0321 05:48:57.441563 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_bb5a0456-b5c5-433a-afde-fe38740e2310/openstackclient/0.log" Mar 21 05:48:57 crc kubenswrapper[4775]: I0321 05:48:57.446935 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_76d205b7-bc2e-4dad-b513-457ff20d67e1/galera/0.log" Mar 21 05:48:57 crc kubenswrapper[4775]: I0321 05:48:57.695550 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-ckh8b_9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8/openstack-network-exporter/0.log" Mar 21 05:48:57 crc kubenswrapper[4775]: I0321 05:48:57.788934 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-nmtjx_8a8e948c-2978-40c8-961b-1b010f7ea920/ovn-controller/0.log" Mar 21 05:48:58 crc kubenswrapper[4775]: I0321 05:48:58.015640 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-frhpj_88a367f7-4951-4e7c-889f-d147676654f8/ovsdb-server-init/0.log" Mar 21 05:48:58 crc kubenswrapper[4775]: I0321 05:48:58.233301 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-frhpj_88a367f7-4951-4e7c-889f-d147676654f8/ovsdb-server-init/0.log" Mar 21 05:48:58 crc kubenswrapper[4775]: I0321 05:48:58.243792 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-frhpj_88a367f7-4951-4e7c-889f-d147676654f8/ovs-vswitchd/0.log" Mar 21 05:48:58 crc kubenswrapper[4775]: I0321 05:48:58.247404 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-frhpj_88a367f7-4951-4e7c-889f-d147676654f8/ovsdb-server/0.log" Mar 21 05:48:58 crc kubenswrapper[4775]: I0321 05:48:58.508381 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_334a5c95-becc-4389-bb6f-50e5957cded6/openstack-network-exporter/0.log" Mar 21 05:48:58 crc kubenswrapper[4775]: I0321 05:48:58.530942 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-45kkj_3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:48:58 crc kubenswrapper[4775]: I0321 05:48:58.557522 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_334a5c95-becc-4389-bb6f-50e5957cded6/ovn-northd/0.log" Mar 21 05:48:58 crc kubenswrapper[4775]: I0321 05:48:58.768994 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e/openstack-network-exporter/0.log" Mar 21 05:48:58 crc kubenswrapper[4775]: I0321 05:48:58.839207 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e/ovsdbserver-nb/0.log" Mar 21 05:48:59 crc kubenswrapper[4775]: I0321 05:48:59.172478 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_93886182-fca2-42a9-a134-2243c7c7073d/openstack-network-exporter/0.log" Mar 21 05:48:59 crc kubenswrapper[4775]: I0321 05:48:59.207587 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_93886182-fca2-42a9-a134-2243c7c7073d/ovsdbserver-sb/0.log" Mar 21 05:48:59 crc kubenswrapper[4775]: I0321 05:48:59.383910 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-58948d8bb4-rcw89_279dff90-9d39-418a-b5e7-00333a376d16/placement-api/0.log" Mar 21 05:48:59 crc kubenswrapper[4775]: I0321 05:48:59.478964 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-58948d8bb4-rcw89_279dff90-9d39-418a-b5e7-00333a376d16/placement-log/0.log" Mar 21 05:48:59 crc kubenswrapper[4775]: I0321 05:48:59.564741 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c95486b5-f2ad-4098-912d-6749b329824b/setup-container/0.log" Mar 21 05:48:59 crc kubenswrapper[4775]: I0321 05:48:59.731046 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c95486b5-f2ad-4098-912d-6749b329824b/setup-container/0.log" Mar 21 05:48:59 crc kubenswrapper[4775]: I0321 05:48:59.800011 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c95486b5-f2ad-4098-912d-6749b329824b/rabbitmq/0.log" Mar 21 05:48:59 crc kubenswrapper[4775]: I0321 05:48:59.874012 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e5e83941-a38d-4ee9-b967-1dac69c5a55b/setup-container/0.log" Mar 21 05:49:00 crc kubenswrapper[4775]: I0321 05:49:00.068847 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e5e83941-a38d-4ee9-b967-1dac69c5a55b/setup-container/0.log" Mar 21 05:49:00 crc kubenswrapper[4775]: I0321 05:49:00.100806 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e5e83941-a38d-4ee9-b967-1dac69c5a55b/rabbitmq/0.log" Mar 21 05:49:00 crc kubenswrapper[4775]: I0321 05:49:00.124039 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6_034f630d-d6d6-41f0-8df6-e5db37b778f3/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:49:00 crc kubenswrapper[4775]: I0321 05:49:00.320314 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-qgpzc_7c88e417-9ede-41d8-8337-79620ceb7798/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:49:00 crc kubenswrapper[4775]: I0321 05:49:00.405993 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m_f41367b2-433d-48f7-af75-575be4b318fc/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:49:00 crc kubenswrapper[4775]: I0321 05:49:00.617921 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-wtk52_8cf3d7cc-425b-4d40-a26c-a88d2b210c0a/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:49:00 crc kubenswrapper[4775]: I0321 05:49:00.635504 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-swg9k_5f45d376-4f59-4584-a545-16d4ff066232/ssh-known-hosts-edpm-deployment/0.log" Mar 21 05:49:00 crc kubenswrapper[4775]: I0321 05:49:00.903290 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-f9b88fb79-vclnv_77432545-f22c-453a-b6a7-7c932712efa9/proxy-server/0.log" Mar 21 05:49:00 crc kubenswrapper[4775]: I0321 05:49:00.947066 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-f9b88fb79-vclnv_77432545-f22c-453a-b6a7-7c932712efa9/proxy-httpd/0.log" Mar 21 05:49:01 crc kubenswrapper[4775]: I0321 05:49:01.030619 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-kzll6_9e521c27-9d67-47bc-b6ac-74fabb543d3f/swift-ring-rebalance/0.log" Mar 21 05:49:01 crc kubenswrapper[4775]: I0321 05:49:01.205157 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/account-auditor/0.log" Mar 21 05:49:01 crc kubenswrapper[4775]: I0321 05:49:01.251183 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/account-reaper/0.log" Mar 21 05:49:01 crc kubenswrapper[4775]: I0321 05:49:01.287061 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/account-replicator/0.log" Mar 21 05:49:01 crc kubenswrapper[4775]: I0321 05:49:01.460562 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/container-replicator/0.log" Mar 21 05:49:01 crc kubenswrapper[4775]: I0321 05:49:01.487453 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/container-auditor/0.log" Mar 21 05:49:01 crc kubenswrapper[4775]: I0321 05:49:01.499398 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/container-server/0.log" Mar 21 05:49:01 crc kubenswrapper[4775]: I0321 05:49:01.505157 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/account-server/0.log" Mar 21 05:49:01 crc kubenswrapper[4775]: I0321 05:49:01.717765 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/object-expirer/0.log" Mar 21 05:49:01 crc kubenswrapper[4775]: I0321 05:49:01.727712 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/container-updater/0.log" Mar 21 05:49:01 crc kubenswrapper[4775]: I0321 05:49:01.734905 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/object-auditor/0.log" Mar 21 05:49:01 crc kubenswrapper[4775]: I0321 05:49:01.766479 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/object-replicator/0.log" Mar 21 05:49:01 crc kubenswrapper[4775]: I0321 05:49:01.945833 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/object-server/0.log" Mar 21 05:49:01 crc kubenswrapper[4775]: I0321 05:49:01.949048 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/object-updater/0.log" Mar 21 05:49:01 crc kubenswrapper[4775]: I0321 05:49:01.969013 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/rsync/0.log" Mar 21 05:49:02 crc kubenswrapper[4775]: I0321 05:49:02.005800 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/swift-recon-cron/0.log" Mar 21 05:49:02 crc kubenswrapper[4775]: I0321 05:49:02.257864 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_1c832898-838d-423d-8ad8-512c5ee5706c/tempest-tests-tempest-tests-runner/0.log" Mar 21 05:49:02 crc kubenswrapper[4775]: I0321 05:49:02.465605 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_2c99faa4-db71-4a05-a018-9c382f33f55e/test-operator-logs-container/0.log" Mar 21 05:49:02 crc kubenswrapper[4775]: I0321 05:49:02.481738 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:49:02 crc kubenswrapper[4775]: I0321 05:49:02.481805 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:49:02 crc kubenswrapper[4775]: I0321 05:49:02.481858 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 05:49:02 crc kubenswrapper[4775]: I0321 05:49:02.483247 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2"} pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 05:49:02 crc kubenswrapper[4775]: I0321 05:49:02.483309 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" containerID="cri-o://fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" gracePeriod=600 Mar 21 05:49:02 crc kubenswrapper[4775]: I0321 05:49:02.626192 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg_bccbefa9-966d-44b9-bd8f-bb566649b315/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:49:02 crc kubenswrapper[4775]: E0321 05:49:02.679702 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:49:02 crc kubenswrapper[4775]: I0321 05:49:02.755676 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-mfft7_268b27f0-a217-459e-9502-7b522ca6fe2c/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:49:03 crc kubenswrapper[4775]: I0321 05:49:03.171188 4775 generic.go:334] "Generic (PLEG): container finished" podID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" exitCode=0 Mar 21 05:49:03 crc kubenswrapper[4775]: I0321 05:49:03.171224 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerDied","Data":"fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2"} Mar 21 05:49:03 crc kubenswrapper[4775]: I0321 05:49:03.171273 4775 scope.go:117] "RemoveContainer" containerID="27384562242ad4f001715c32b1c4d680359f25816c963e4f026f07e0e28e85a2" Mar 21 05:49:03 crc kubenswrapper[4775]: I0321 05:49:03.172107 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:49:03 crc kubenswrapper[4775]: E0321 05:49:03.172505 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:49:11 crc kubenswrapper[4775]: I0321 05:49:11.743425 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_2a8305a2-5178-437f-a896-314b34fa595e/memcached/0.log" Mar 21 05:49:14 crc kubenswrapper[4775]: I0321 05:49:14.661718 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:49:14 crc kubenswrapper[4775]: E0321 05:49:14.662554 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:49:29 crc kubenswrapper[4775]: I0321 05:49:29.661294 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:49:29 crc kubenswrapper[4775]: E0321 05:49:29.661887 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:49:32 crc kubenswrapper[4775]: I0321 05:49:32.093606 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt_a8157e0e-bf83-4dbd-af86-d09d14e6e1b1/util/0.log" Mar 21 05:49:32 crc kubenswrapper[4775]: I0321 05:49:32.226902 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt_a8157e0e-bf83-4dbd-af86-d09d14e6e1b1/util/0.log" Mar 21 05:49:32 crc kubenswrapper[4775]: I0321 05:49:32.257350 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt_a8157e0e-bf83-4dbd-af86-d09d14e6e1b1/pull/0.log" Mar 21 05:49:32 crc kubenswrapper[4775]: I0321 05:49:32.305941 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt_a8157e0e-bf83-4dbd-af86-d09d14e6e1b1/pull/0.log" Mar 21 05:49:32 crc kubenswrapper[4775]: I0321 05:49:32.511747 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt_a8157e0e-bf83-4dbd-af86-d09d14e6e1b1/extract/0.log" Mar 21 05:49:32 crc kubenswrapper[4775]: I0321 05:49:32.530268 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt_a8157e0e-bf83-4dbd-af86-d09d14e6e1b1/util/0.log" Mar 21 05:49:32 crc kubenswrapper[4775]: I0321 05:49:32.537819 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt_a8157e0e-bf83-4dbd-af86-d09d14e6e1b1/pull/0.log" Mar 21 05:49:32 crc kubenswrapper[4775]: I0321 05:49:32.766456 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-59bc569d95-twhxx_94e1507b-be6c-4ecf-99e5-2bdcd2cc0cef/manager/0.log" Mar 21 05:49:33 crc kubenswrapper[4775]: I0321 05:49:33.085081 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-588d4d986b-lxvtw_87dcea67-7f65-46a6-996b-3985bf1b5171/manager/0.log" Mar 21 05:49:33 crc kubenswrapper[4775]: I0321 05:49:33.282009 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-79df6bcc97-lzvsq_80932361-6406-48dd-9e4b-4e9c27813f68/manager/0.log" Mar 21 05:49:33 crc kubenswrapper[4775]: I0321 05:49:33.458669 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-67dd5f86f5-m4gqz_0e83601c-758c-4f12-b745-bb68b0c4904f/manager/0.log" Mar 21 05:49:33 crc kubenswrapper[4775]: I0321 05:49:33.635609 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-8464cc45fb-7bdbc_01ed348d-8a8a-4717-ba0d-1944b3f1c081/manager/0.log" Mar 21 05:49:33 crc kubenswrapper[4775]: I0321 05:49:33.681738 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-8d58dc466-dn22m_9fe71acc-7d35-4d4b-ac69-e193d3f39028/manager/0.log" Mar 21 05:49:33 crc kubenswrapper[4775]: I0321 05:49:33.957069 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-65f65cc49c-2mgp8_0a66456f-7860-4dc1-9c1c-0db69ddcc800/manager/0.log" Mar 21 05:49:34 crc kubenswrapper[4775]: I0321 05:49:34.123839 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6f787dddc9-9g8pv_f00c8c4b-874f-45ec-8a1a-e0834b3fc252/manager/0.log" Mar 21 05:49:34 crc kubenswrapper[4775]: I0321 05:49:34.284335 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-55f864c847-jq88h_20c78c73-daf5-481e-a4ac-62de73b5969e/manager/0.log" Mar 21 05:49:34 crc kubenswrapper[4775]: I0321 05:49:34.331290 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-768b96df4c-c7pjw_49a3c6c7-9e86-495d-8d1a-486d6bfbbbdd/manager/0.log" Mar 21 05:49:34 crc kubenswrapper[4775]: I0321 05:49:34.571959 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67ccfc9778-hf688_ec38d53e-6fe4-41b5-8548-e49fadd9d6bf/manager/0.log" Mar 21 05:49:34 crc kubenswrapper[4775]: I0321 05:49:34.581914 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-767865f676-4tbvg_9bfe7d25-53ea-484d-a481-0ea04ee2b8a8/manager/0.log" Mar 21 05:49:34 crc kubenswrapper[4775]: I0321 05:49:34.803016 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5b9f45d989-67scw_745c79b1-1bcf-4c0f-82ee-a26cbba46d48/manager/0.log" Mar 21 05:49:34 crc kubenswrapper[4775]: I0321 05:49:34.921295 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5d488d59fb-tss7r_02b6af47-2c06-480b-a838-2d742efa1045/manager/0.log" Mar 21 05:49:35 crc kubenswrapper[4775]: I0321 05:49:35.033474 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-89d64c458-rwq59_3035739a-202f-4794-bb4f-ae2342a96441/manager/0.log" Mar 21 05:49:35 crc kubenswrapper[4775]: I0321 05:49:35.137515 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-85fcfb8fbb-q2k4k_899f7d20-7208-419e-b0f8-36c7fbf2e841/operator/0.log" Mar 21 05:49:35 crc kubenswrapper[4775]: I0321 05:49:35.384014 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-hvtkg_8c7426e8-8cec-4c84-8810-03a091d87cd9/registry-server/0.log" Mar 21 05:49:35 crc kubenswrapper[4775]: I0321 05:49:35.604953 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-884679f54-ctm9h_6eeb04ad-7251-488c-bd52-b2f14f6fb68b/manager/0.log" Mar 21 05:49:35 crc kubenswrapper[4775]: I0321 05:49:35.861284 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5784578c99-jj4pt_898b32c5-9f21-4fba-90c5-a333f36addf2/manager/0.log" Mar 21 05:49:36 crc kubenswrapper[4775]: I0321 05:49:36.078596 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-jvglf_95a44b12-e027-400d-b257-99f2012251d8/operator/0.log" Mar 21 05:49:36 crc kubenswrapper[4775]: I0321 05:49:36.251958 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-c674c5965-lqmgv_5968f1d9-f4e0-4c67-923e-2494e15c4088/manager/0.log" Mar 21 05:49:36 crc kubenswrapper[4775]: I0321 05:49:36.414230 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-d6b694c5-9s82h_d1dbd80a-0782-4035-a263-b52a90f6ee0e/manager/0.log" Mar 21 05:49:36 crc kubenswrapper[4775]: I0321 05:49:36.518541 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-l59gx_9cac78ed-6325-4649-bb05-a1518ae692e9/manager/0.log" Mar 21 05:49:36 crc kubenswrapper[4775]: I0321 05:49:36.544935 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-65746ff4dc-hg4rq_907f0cdf-2d87-4d09-97af-5591d061b4f6/manager/0.log" Mar 21 05:49:36 crc kubenswrapper[4775]: I0321 05:49:36.665471 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6c4d75f7f9-gdrc5_3c9f18bd-def6-45ff-a92b-25c6f40d6bb5/manager/0.log" Mar 21 05:49:40 crc kubenswrapper[4775]: I0321 05:49:40.661604 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:49:40 crc kubenswrapper[4775]: E0321 05:49:40.662245 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:49:55 crc kubenswrapper[4775]: I0321 05:49:55.662397 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:49:55 crc kubenswrapper[4775]: E0321 05:49:55.663253 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:49:57 crc kubenswrapper[4775]: I0321 05:49:57.226140 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-d7xpn_e5224539-6d29-4bc3-9656-4665eb287e28/control-plane-machine-set-operator/0.log" Mar 21 05:49:57 crc kubenswrapper[4775]: I0321 05:49:57.429022 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5rnj6_fe3df1e1-4c22-48df-aaea-469c864f0310/kube-rbac-proxy/0.log" Mar 21 05:49:57 crc kubenswrapper[4775]: I0321 05:49:57.468523 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5rnj6_fe3df1e1-4c22-48df-aaea-469c864f0310/machine-api-operator/0.log" Mar 21 05:50:00 crc kubenswrapper[4775]: I0321 05:50:00.144015 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567870-hdgmt"] Mar 21 05:50:00 crc kubenswrapper[4775]: E0321 05:50:00.144980 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c04654f3-0f1c-4883-8385-4149f0076d44" containerName="container-00" Mar 21 05:50:00 crc kubenswrapper[4775]: I0321 05:50:00.144999 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c04654f3-0f1c-4883-8385-4149f0076d44" containerName="container-00" Mar 21 05:50:00 crc kubenswrapper[4775]: I0321 05:50:00.145298 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c04654f3-0f1c-4883-8385-4149f0076d44" containerName="container-00" Mar 21 05:50:00 crc kubenswrapper[4775]: I0321 05:50:00.146174 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567870-hdgmt" Mar 21 05:50:00 crc kubenswrapper[4775]: I0321 05:50:00.148422 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:50:00 crc kubenswrapper[4775]: I0321 05:50:00.148498 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:50:00 crc kubenswrapper[4775]: I0321 05:50:00.149483 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:50:00 crc kubenswrapper[4775]: I0321 05:50:00.152880 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567870-hdgmt"] Mar 21 05:50:00 crc kubenswrapper[4775]: I0321 05:50:00.295953 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kckln\" (UniqueName: \"kubernetes.io/projected/3b54d80b-a103-474c-8214-44ec9b595a43-kube-api-access-kckln\") pod \"auto-csr-approver-29567870-hdgmt\" (UID: \"3b54d80b-a103-474c-8214-44ec9b595a43\") " pod="openshift-infra/auto-csr-approver-29567870-hdgmt" Mar 21 05:50:00 crc kubenswrapper[4775]: I0321 05:50:00.398349 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kckln\" (UniqueName: \"kubernetes.io/projected/3b54d80b-a103-474c-8214-44ec9b595a43-kube-api-access-kckln\") pod \"auto-csr-approver-29567870-hdgmt\" (UID: \"3b54d80b-a103-474c-8214-44ec9b595a43\") " pod="openshift-infra/auto-csr-approver-29567870-hdgmt" Mar 21 05:50:00 crc kubenswrapper[4775]: I0321 05:50:00.427315 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kckln\" (UniqueName: \"kubernetes.io/projected/3b54d80b-a103-474c-8214-44ec9b595a43-kube-api-access-kckln\") pod \"auto-csr-approver-29567870-hdgmt\" (UID: \"3b54d80b-a103-474c-8214-44ec9b595a43\") " pod="openshift-infra/auto-csr-approver-29567870-hdgmt" Mar 21 05:50:00 crc kubenswrapper[4775]: I0321 05:50:00.517798 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567870-hdgmt" Mar 21 05:50:01 crc kubenswrapper[4775]: I0321 05:50:01.002337 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567870-hdgmt"] Mar 21 05:50:01 crc kubenswrapper[4775]: I0321 05:50:01.005431 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 05:50:01 crc kubenswrapper[4775]: I0321 05:50:01.632011 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567870-hdgmt" event={"ID":"3b54d80b-a103-474c-8214-44ec9b595a43","Type":"ContainerStarted","Data":"cda58da76eb41a6cdea286216edcfc30a8f820e91ef784de81a6113e9dc45d9b"} Mar 21 05:50:02 crc kubenswrapper[4775]: I0321 05:50:02.641503 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567870-hdgmt" event={"ID":"3b54d80b-a103-474c-8214-44ec9b595a43","Type":"ContainerStarted","Data":"21e7854b61e199821209ba3f9487dda8cf82083673134efff058e8fd96d17f3f"} Mar 21 05:50:03 crc kubenswrapper[4775]: I0321 05:50:03.653153 4775 generic.go:334] "Generic (PLEG): container finished" podID="3b54d80b-a103-474c-8214-44ec9b595a43" containerID="21e7854b61e199821209ba3f9487dda8cf82083673134efff058e8fd96d17f3f" exitCode=0 Mar 21 05:50:03 crc kubenswrapper[4775]: I0321 05:50:03.653231 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567870-hdgmt" event={"ID":"3b54d80b-a103-474c-8214-44ec9b595a43","Type":"ContainerDied","Data":"21e7854b61e199821209ba3f9487dda8cf82083673134efff058e8fd96d17f3f"} Mar 21 05:50:05 crc kubenswrapper[4775]: I0321 05:50:05.075836 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567870-hdgmt" Mar 21 05:50:05 crc kubenswrapper[4775]: I0321 05:50:05.216928 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kckln\" (UniqueName: \"kubernetes.io/projected/3b54d80b-a103-474c-8214-44ec9b595a43-kube-api-access-kckln\") pod \"3b54d80b-a103-474c-8214-44ec9b595a43\" (UID: \"3b54d80b-a103-474c-8214-44ec9b595a43\") " Mar 21 05:50:05 crc kubenswrapper[4775]: I0321 05:50:05.222788 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b54d80b-a103-474c-8214-44ec9b595a43-kube-api-access-kckln" (OuterVolumeSpecName: "kube-api-access-kckln") pod "3b54d80b-a103-474c-8214-44ec9b595a43" (UID: "3b54d80b-a103-474c-8214-44ec9b595a43"). InnerVolumeSpecName "kube-api-access-kckln". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:50:05 crc kubenswrapper[4775]: I0321 05:50:05.319836 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kckln\" (UniqueName: \"kubernetes.io/projected/3b54d80b-a103-474c-8214-44ec9b595a43-kube-api-access-kckln\") on node \"crc\" DevicePath \"\"" Mar 21 05:50:05 crc kubenswrapper[4775]: I0321 05:50:05.673022 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567870-hdgmt" event={"ID":"3b54d80b-a103-474c-8214-44ec9b595a43","Type":"ContainerDied","Data":"cda58da76eb41a6cdea286216edcfc30a8f820e91ef784de81a6113e9dc45d9b"} Mar 21 05:50:05 crc kubenswrapper[4775]: I0321 05:50:05.673073 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cda58da76eb41a6cdea286216edcfc30a8f820e91ef784de81a6113e9dc45d9b" Mar 21 05:50:05 crc kubenswrapper[4775]: I0321 05:50:05.673669 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567870-hdgmt" Mar 21 05:50:05 crc kubenswrapper[4775]: I0321 05:50:05.734590 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567864-zsztj"] Mar 21 05:50:05 crc kubenswrapper[4775]: I0321 05:50:05.745087 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567864-zsztj"] Mar 21 05:50:07 crc kubenswrapper[4775]: I0321 05:50:07.671533 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3565ae50-5b89-4775-a8c2-45209b86619f" path="/var/lib/kubelet/pods/3565ae50-5b89-4775-a8c2-45209b86619f/volumes" Mar 21 05:50:09 crc kubenswrapper[4775]: I0321 05:50:09.662028 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:50:09 crc kubenswrapper[4775]: E0321 05:50:09.662795 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:50:10 crc kubenswrapper[4775]: I0321 05:50:10.469722 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-c2lsw_4fdc8b75-b0a1-4ed3-9eee-6ee726dd0fbe/cert-manager-controller/0.log" Mar 21 05:50:10 crc kubenswrapper[4775]: I0321 05:50:10.684412 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-ngckr_172b2006-3394-469a-be7f-1b66d020fd45/cert-manager-cainjector/0.log" Mar 21 05:50:10 crc kubenswrapper[4775]: I0321 05:50:10.736971 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-ms9hv_670f734f-e215-441e-9b56-7251bc7f2484/cert-manager-webhook/0.log" Mar 21 05:50:22 crc kubenswrapper[4775]: I0321 05:50:22.661485 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:50:22 crc kubenswrapper[4775]: E0321 05:50:22.662458 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:50:23 crc kubenswrapper[4775]: I0321 05:50:23.735456 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-8x7x2_b0fbab95-1c88-40a3-8ccb-58bca74c8f3c/nmstate-console-plugin/0.log" Mar 21 05:50:23 crc kubenswrapper[4775]: I0321 05:50:23.924302 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-6l74f_d03e5939-1625-4597-ad3b-9edf8e8075f5/nmstate-handler/0.log" Mar 21 05:50:23 crc kubenswrapper[4775]: I0321 05:50:23.990907 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-626kh_409422f2-717f-4f82-8dae-fe01dfda7083/kube-rbac-proxy/0.log" Mar 21 05:50:24 crc kubenswrapper[4775]: I0321 05:50:24.059703 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-626kh_409422f2-717f-4f82-8dae-fe01dfda7083/nmstate-metrics/0.log" Mar 21 05:50:24 crc kubenswrapper[4775]: I0321 05:50:24.187061 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-xljlp_f4a0a79a-5b67-44f8-9ef5-304530d5e764/nmstate-operator/0.log" Mar 21 05:50:24 crc kubenswrapper[4775]: I0321 05:50:24.259318 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-rptn7_944b76e5-c8c5-4cba-9df2-9e9b87b540a8/nmstate-webhook/0.log" Mar 21 05:50:33 crc kubenswrapper[4775]: I0321 05:50:33.662016 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:50:33 crc kubenswrapper[4775]: E0321 05:50:33.662809 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:50:48 crc kubenswrapper[4775]: I0321 05:50:48.661578 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:50:48 crc kubenswrapper[4775]: E0321 05:50:48.662463 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:50:49 crc kubenswrapper[4775]: I0321 05:50:49.252413 4775 scope.go:117] "RemoveContainer" containerID="dbc52e5cee44a0750d2409abd45a8584ff150c5de4624d45b211f61e09e87efd" Mar 21 05:50:50 crc kubenswrapper[4775]: I0321 05:50:50.327530 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-qq976_ac0be1f3-95f0-40a4-9a94-c74cdaad9590/kube-rbac-proxy/0.log" Mar 21 05:50:50 crc kubenswrapper[4775]: I0321 05:50:50.444085 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-qq976_ac0be1f3-95f0-40a4-9a94-c74cdaad9590/controller/0.log" Mar 21 05:50:50 crc kubenswrapper[4775]: I0321 05:50:50.559571 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-frr-files/0.log" Mar 21 05:50:50 crc kubenswrapper[4775]: I0321 05:50:50.786712 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-metrics/0.log" Mar 21 05:50:50 crc kubenswrapper[4775]: I0321 05:50:50.794817 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-reloader/0.log" Mar 21 05:50:50 crc kubenswrapper[4775]: I0321 05:50:50.823083 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-reloader/0.log" Mar 21 05:50:50 crc kubenswrapper[4775]: I0321 05:50:50.830707 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-frr-files/0.log" Mar 21 05:50:51 crc kubenswrapper[4775]: I0321 05:50:51.022324 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-reloader/0.log" Mar 21 05:50:51 crc kubenswrapper[4775]: I0321 05:50:51.077079 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-frr-files/0.log" Mar 21 05:50:51 crc kubenswrapper[4775]: I0321 05:50:51.100477 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-metrics/0.log" Mar 21 05:50:51 crc kubenswrapper[4775]: I0321 05:50:51.102315 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-metrics/0.log" Mar 21 05:50:51 crc kubenswrapper[4775]: I0321 05:50:51.368788 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-frr-files/0.log" Mar 21 05:50:51 crc kubenswrapper[4775]: I0321 05:50:51.389628 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-metrics/0.log" Mar 21 05:50:51 crc kubenswrapper[4775]: I0321 05:50:51.400762 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-reloader/0.log" Mar 21 05:50:51 crc kubenswrapper[4775]: I0321 05:50:51.404569 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/controller/0.log" Mar 21 05:50:51 crc kubenswrapper[4775]: I0321 05:50:51.636960 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/frr-metrics/0.log" Mar 21 05:50:51 crc kubenswrapper[4775]: I0321 05:50:51.651689 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/kube-rbac-proxy-frr/0.log" Mar 21 05:50:51 crc kubenswrapper[4775]: I0321 05:50:51.692415 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/kube-rbac-proxy/0.log" Mar 21 05:50:51 crc kubenswrapper[4775]: I0321 05:50:51.871223 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/reloader/0.log" Mar 21 05:50:51 crc kubenswrapper[4775]: I0321 05:50:51.940896 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-7jsnx_e5808d9b-074a-4948-8283-fdfea77c63bc/frr-k8s-webhook-server/0.log" Mar 21 05:50:52 crc kubenswrapper[4775]: I0321 05:50:52.222231 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6695f56dbb-f6tqn_38d80f78-fa33-49b7-99c2-62d50d1c011b/manager/0.log" Mar 21 05:50:52 crc kubenswrapper[4775]: I0321 05:50:52.572916 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-559bfcf5c-qqsvn_8c52832d-1aee-4eac-b625-24110b985402/webhook-server/0.log" Mar 21 05:50:52 crc kubenswrapper[4775]: I0321 05:50:52.744513 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cpw6m_1d9349cc-e186-40c5-bb71-c176ff4f0a0d/kube-rbac-proxy/0.log" Mar 21 05:50:52 crc kubenswrapper[4775]: I0321 05:50:52.857453 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/frr/0.log" Mar 21 05:50:53 crc kubenswrapper[4775]: I0321 05:50:53.206712 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cpw6m_1d9349cc-e186-40c5-bb71-c176ff4f0a0d/speaker/0.log" Mar 21 05:51:03 crc kubenswrapper[4775]: I0321 05:51:03.661502 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:51:03 crc kubenswrapper[4775]: E0321 05:51:03.662339 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:51:06 crc kubenswrapper[4775]: I0321 05:51:06.789886 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp_b2ca57eb-edac-4db7-bfa4-3198d80daf99/util/0.log" Mar 21 05:51:06 crc kubenswrapper[4775]: I0321 05:51:06.998202 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp_b2ca57eb-edac-4db7-bfa4-3198d80daf99/util/0.log" Mar 21 05:51:07 crc kubenswrapper[4775]: I0321 05:51:07.018981 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp_b2ca57eb-edac-4db7-bfa4-3198d80daf99/pull/0.log" Mar 21 05:51:07 crc kubenswrapper[4775]: I0321 05:51:07.024482 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp_b2ca57eb-edac-4db7-bfa4-3198d80daf99/pull/0.log" Mar 21 05:51:07 crc kubenswrapper[4775]: I0321 05:51:07.175411 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp_b2ca57eb-edac-4db7-bfa4-3198d80daf99/util/0.log" Mar 21 05:51:07 crc kubenswrapper[4775]: I0321 05:51:07.208105 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp_b2ca57eb-edac-4db7-bfa4-3198d80daf99/extract/0.log" Mar 21 05:51:07 crc kubenswrapper[4775]: I0321 05:51:07.246037 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp_b2ca57eb-edac-4db7-bfa4-3198d80daf99/pull/0.log" Mar 21 05:51:07 crc kubenswrapper[4775]: I0321 05:51:07.344142 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s_6252c43a-d149-46f7-ac6c-263c34980fe2/util/0.log" Mar 21 05:51:07 crc kubenswrapper[4775]: I0321 05:51:07.517983 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s_6252c43a-d149-46f7-ac6c-263c34980fe2/util/0.log" Mar 21 05:51:07 crc kubenswrapper[4775]: I0321 05:51:07.533065 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s_6252c43a-d149-46f7-ac6c-263c34980fe2/pull/0.log" Mar 21 05:51:07 crc kubenswrapper[4775]: I0321 05:51:07.538750 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s_6252c43a-d149-46f7-ac6c-263c34980fe2/pull/0.log" Mar 21 05:51:07 crc kubenswrapper[4775]: I0321 05:51:07.729237 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s_6252c43a-d149-46f7-ac6c-263c34980fe2/pull/0.log" Mar 21 05:51:07 crc kubenswrapper[4775]: I0321 05:51:07.746748 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s_6252c43a-d149-46f7-ac6c-263c34980fe2/util/0.log" Mar 21 05:51:07 crc kubenswrapper[4775]: I0321 05:51:07.765424 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s_6252c43a-d149-46f7-ac6c-263c34980fe2/extract/0.log" Mar 21 05:51:07 crc kubenswrapper[4775]: I0321 05:51:07.919085 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77cp9_08558c65-4599-4c64-bb0f-f18f94cecdec/extract-utilities/0.log" Mar 21 05:51:08 crc kubenswrapper[4775]: I0321 05:51:08.100741 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77cp9_08558c65-4599-4c64-bb0f-f18f94cecdec/extract-content/0.log" Mar 21 05:51:08 crc kubenswrapper[4775]: I0321 05:51:08.121669 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77cp9_08558c65-4599-4c64-bb0f-f18f94cecdec/extract-utilities/0.log" Mar 21 05:51:08 crc kubenswrapper[4775]: I0321 05:51:08.123020 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77cp9_08558c65-4599-4c64-bb0f-f18f94cecdec/extract-content/0.log" Mar 21 05:51:08 crc kubenswrapper[4775]: I0321 05:51:08.334055 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77cp9_08558c65-4599-4c64-bb0f-f18f94cecdec/extract-content/0.log" Mar 21 05:51:08 crc kubenswrapper[4775]: I0321 05:51:08.352129 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77cp9_08558c65-4599-4c64-bb0f-f18f94cecdec/extract-utilities/0.log" Mar 21 05:51:08 crc kubenswrapper[4775]: I0321 05:51:08.566760 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x59zn_e8513867-3729-4d93-b8ca-45ecb69c50e6/extract-utilities/0.log" Mar 21 05:51:08 crc kubenswrapper[4775]: I0321 05:51:08.841401 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x59zn_e8513867-3729-4d93-b8ca-45ecb69c50e6/extract-content/0.log" Mar 21 05:51:08 crc kubenswrapper[4775]: I0321 05:51:08.883077 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x59zn_e8513867-3729-4d93-b8ca-45ecb69c50e6/extract-utilities/0.log" Mar 21 05:51:08 crc kubenswrapper[4775]: I0321 05:51:08.901084 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x59zn_e8513867-3729-4d93-b8ca-45ecb69c50e6/extract-content/0.log" Mar 21 05:51:09 crc kubenswrapper[4775]: I0321 05:51:09.007097 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77cp9_08558c65-4599-4c64-bb0f-f18f94cecdec/registry-server/0.log" Mar 21 05:51:09 crc kubenswrapper[4775]: I0321 05:51:09.084548 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x59zn_e8513867-3729-4d93-b8ca-45ecb69c50e6/extract-content/0.log" Mar 21 05:51:09 crc kubenswrapper[4775]: I0321 05:51:09.087952 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x59zn_e8513867-3729-4d93-b8ca-45ecb69c50e6/extract-utilities/0.log" Mar 21 05:51:09 crc kubenswrapper[4775]: I0321 05:51:09.442408 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-z7fcx_59fec450-4b61-4a15-b1b5-b47dedd649a0/marketplace-operator/0.log" Mar 21 05:51:09 crc kubenswrapper[4775]: I0321 05:51:09.527579 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h4wmk_4a9e7d52-67b3-4a38-a978-6566b1c7870a/extract-utilities/0.log" Mar 21 05:51:09 crc kubenswrapper[4775]: I0321 05:51:09.746811 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h4wmk_4a9e7d52-67b3-4a38-a978-6566b1c7870a/extract-utilities/0.log" Mar 21 05:51:09 crc kubenswrapper[4775]: I0321 05:51:09.770313 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h4wmk_4a9e7d52-67b3-4a38-a978-6566b1c7870a/extract-content/0.log" Mar 21 05:51:09 crc kubenswrapper[4775]: I0321 05:51:09.792010 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h4wmk_4a9e7d52-67b3-4a38-a978-6566b1c7870a/extract-content/0.log" Mar 21 05:51:09 crc kubenswrapper[4775]: I0321 05:51:09.798625 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x59zn_e8513867-3729-4d93-b8ca-45ecb69c50e6/registry-server/0.log" Mar 21 05:51:10 crc kubenswrapper[4775]: I0321 05:51:10.017153 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h4wmk_4a9e7d52-67b3-4a38-a978-6566b1c7870a/extract-content/0.log" Mar 21 05:51:10 crc kubenswrapper[4775]: I0321 05:51:10.019553 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h4wmk_4a9e7d52-67b3-4a38-a978-6566b1c7870a/extract-utilities/0.log" Mar 21 05:51:10 crc kubenswrapper[4775]: I0321 05:51:10.065087 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h4wmk_4a9e7d52-67b3-4a38-a978-6566b1c7870a/registry-server/0.log" Mar 21 05:51:10 crc kubenswrapper[4775]: I0321 05:51:10.203699 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gslhk_373fa4a4-80b8-4b32-a9ee-b272604d3adc/extract-utilities/0.log" Mar 21 05:51:10 crc kubenswrapper[4775]: I0321 05:51:10.388421 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gslhk_373fa4a4-80b8-4b32-a9ee-b272604d3adc/extract-content/0.log" Mar 21 05:51:10 crc kubenswrapper[4775]: I0321 05:51:10.401580 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gslhk_373fa4a4-80b8-4b32-a9ee-b272604d3adc/extract-content/0.log" Mar 21 05:51:10 crc kubenswrapper[4775]: I0321 05:51:10.411495 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gslhk_373fa4a4-80b8-4b32-a9ee-b272604d3adc/extract-utilities/0.log" Mar 21 05:51:10 crc kubenswrapper[4775]: I0321 05:51:10.597776 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gslhk_373fa4a4-80b8-4b32-a9ee-b272604d3adc/extract-utilities/0.log" Mar 21 05:51:10 crc kubenswrapper[4775]: I0321 05:51:10.629830 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gslhk_373fa4a4-80b8-4b32-a9ee-b272604d3adc/extract-content/0.log" Mar 21 05:51:11 crc kubenswrapper[4775]: I0321 05:51:11.733534 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gslhk_373fa4a4-80b8-4b32-a9ee-b272604d3adc/registry-server/0.log" Mar 21 05:51:15 crc kubenswrapper[4775]: I0321 05:51:15.661768 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:51:15 crc kubenswrapper[4775]: E0321 05:51:15.662439 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:51:20 crc kubenswrapper[4775]: I0321 05:51:20.443844 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-snsfc"] Mar 21 05:51:20 crc kubenswrapper[4775]: E0321 05:51:20.444924 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b54d80b-a103-474c-8214-44ec9b595a43" containerName="oc" Mar 21 05:51:20 crc kubenswrapper[4775]: I0321 05:51:20.444937 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b54d80b-a103-474c-8214-44ec9b595a43" containerName="oc" Mar 21 05:51:20 crc kubenswrapper[4775]: I0321 05:51:20.445162 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b54d80b-a103-474c-8214-44ec9b595a43" containerName="oc" Mar 21 05:51:20 crc kubenswrapper[4775]: I0321 05:51:20.446538 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-snsfc" Mar 21 05:51:20 crc kubenswrapper[4775]: I0321 05:51:20.456665 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-snsfc"] Mar 21 05:51:20 crc kubenswrapper[4775]: I0321 05:51:20.517220 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79dd9ac3-231b-4217-a717-a3d127b0c0ca-utilities\") pod \"redhat-operators-snsfc\" (UID: \"79dd9ac3-231b-4217-a717-a3d127b0c0ca\") " pod="openshift-marketplace/redhat-operators-snsfc" Mar 21 05:51:20 crc kubenswrapper[4775]: I0321 05:51:20.517290 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqhpz\" (UniqueName: \"kubernetes.io/projected/79dd9ac3-231b-4217-a717-a3d127b0c0ca-kube-api-access-jqhpz\") pod \"redhat-operators-snsfc\" (UID: \"79dd9ac3-231b-4217-a717-a3d127b0c0ca\") " pod="openshift-marketplace/redhat-operators-snsfc" Mar 21 05:51:20 crc kubenswrapper[4775]: I0321 05:51:20.517449 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79dd9ac3-231b-4217-a717-a3d127b0c0ca-catalog-content\") pod \"redhat-operators-snsfc\" (UID: \"79dd9ac3-231b-4217-a717-a3d127b0c0ca\") " pod="openshift-marketplace/redhat-operators-snsfc" Mar 21 05:51:20 crc kubenswrapper[4775]: I0321 05:51:20.619766 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79dd9ac3-231b-4217-a717-a3d127b0c0ca-catalog-content\") pod \"redhat-operators-snsfc\" (UID: \"79dd9ac3-231b-4217-a717-a3d127b0c0ca\") " pod="openshift-marketplace/redhat-operators-snsfc" Mar 21 05:51:20 crc kubenswrapper[4775]: I0321 05:51:20.619930 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79dd9ac3-231b-4217-a717-a3d127b0c0ca-utilities\") pod \"redhat-operators-snsfc\" (UID: \"79dd9ac3-231b-4217-a717-a3d127b0c0ca\") " pod="openshift-marketplace/redhat-operators-snsfc" Mar 21 05:51:20 crc kubenswrapper[4775]: I0321 05:51:20.619992 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqhpz\" (UniqueName: \"kubernetes.io/projected/79dd9ac3-231b-4217-a717-a3d127b0c0ca-kube-api-access-jqhpz\") pod \"redhat-operators-snsfc\" (UID: \"79dd9ac3-231b-4217-a717-a3d127b0c0ca\") " pod="openshift-marketplace/redhat-operators-snsfc" Mar 21 05:51:20 crc kubenswrapper[4775]: I0321 05:51:20.620363 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79dd9ac3-231b-4217-a717-a3d127b0c0ca-catalog-content\") pod \"redhat-operators-snsfc\" (UID: \"79dd9ac3-231b-4217-a717-a3d127b0c0ca\") " pod="openshift-marketplace/redhat-operators-snsfc" Mar 21 05:51:20 crc kubenswrapper[4775]: I0321 05:51:20.620680 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79dd9ac3-231b-4217-a717-a3d127b0c0ca-utilities\") pod \"redhat-operators-snsfc\" (UID: \"79dd9ac3-231b-4217-a717-a3d127b0c0ca\") " pod="openshift-marketplace/redhat-operators-snsfc" Mar 21 05:51:20 crc kubenswrapper[4775]: I0321 05:51:20.641207 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqhpz\" (UniqueName: \"kubernetes.io/projected/79dd9ac3-231b-4217-a717-a3d127b0c0ca-kube-api-access-jqhpz\") pod \"redhat-operators-snsfc\" (UID: \"79dd9ac3-231b-4217-a717-a3d127b0c0ca\") " pod="openshift-marketplace/redhat-operators-snsfc" Mar 21 05:51:20 crc kubenswrapper[4775]: I0321 05:51:20.766382 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-snsfc" Mar 21 05:51:21 crc kubenswrapper[4775]: I0321 05:51:21.314699 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-snsfc"] Mar 21 05:51:21 crc kubenswrapper[4775]: I0321 05:51:21.384047 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snsfc" event={"ID":"79dd9ac3-231b-4217-a717-a3d127b0c0ca","Type":"ContainerStarted","Data":"60650042fc6959477421db51b293d07bc331592b53cc199339128f46e38e52de"} Mar 21 05:51:22 crc kubenswrapper[4775]: I0321 05:51:22.393570 4775 generic.go:334] "Generic (PLEG): container finished" podID="79dd9ac3-231b-4217-a717-a3d127b0c0ca" containerID="a695039e53bef0f3edbd44ff68e271754c9a8e6448d909abcaca31bcc1cb9ed6" exitCode=0 Mar 21 05:51:22 crc kubenswrapper[4775]: I0321 05:51:22.393781 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snsfc" event={"ID":"79dd9ac3-231b-4217-a717-a3d127b0c0ca","Type":"ContainerDied","Data":"a695039e53bef0f3edbd44ff68e271754c9a8e6448d909abcaca31bcc1cb9ed6"} Mar 21 05:51:23 crc kubenswrapper[4775]: I0321 05:51:23.404759 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snsfc" event={"ID":"79dd9ac3-231b-4217-a717-a3d127b0c0ca","Type":"ContainerStarted","Data":"6b761fc8b5585d54f920ab7f638b663e14b509ba0f6869c22d0a1c572cfc5a8d"} Mar 21 05:51:29 crc kubenswrapper[4775]: I0321 05:51:29.500647 4775 generic.go:334] "Generic (PLEG): container finished" podID="79dd9ac3-231b-4217-a717-a3d127b0c0ca" containerID="6b761fc8b5585d54f920ab7f638b663e14b509ba0f6869c22d0a1c572cfc5a8d" exitCode=0 Mar 21 05:51:29 crc kubenswrapper[4775]: I0321 05:51:29.500940 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snsfc" event={"ID":"79dd9ac3-231b-4217-a717-a3d127b0c0ca","Type":"ContainerDied","Data":"6b761fc8b5585d54f920ab7f638b663e14b509ba0f6869c22d0a1c572cfc5a8d"} Mar 21 05:51:30 crc kubenswrapper[4775]: I0321 05:51:30.661318 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:51:30 crc kubenswrapper[4775]: E0321 05:51:30.661897 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:51:31 crc kubenswrapper[4775]: I0321 05:51:31.521448 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snsfc" event={"ID":"79dd9ac3-231b-4217-a717-a3d127b0c0ca","Type":"ContainerStarted","Data":"a9c141457cd1a8a69f1b8a93fbbbea5921c7cfed496f1b33f8197ea1b1d27341"} Mar 21 05:51:40 crc kubenswrapper[4775]: I0321 05:51:40.768488 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-snsfc" Mar 21 05:51:40 crc kubenswrapper[4775]: I0321 05:51:40.770130 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-snsfc" Mar 21 05:51:41 crc kubenswrapper[4775]: I0321 05:51:41.661688 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:51:41 crc kubenswrapper[4775]: E0321 05:51:41.662226 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:51:41 crc kubenswrapper[4775]: I0321 05:51:41.831871 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-snsfc" podUID="79dd9ac3-231b-4217-a717-a3d127b0c0ca" containerName="registry-server" probeResult="failure" output=< Mar 21 05:51:41 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Mar 21 05:51:41 crc kubenswrapper[4775]: > Mar 21 05:51:51 crc kubenswrapper[4775]: I0321 05:51:51.823317 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-snsfc" podUID="79dd9ac3-231b-4217-a717-a3d127b0c0ca" containerName="registry-server" probeResult="failure" output=< Mar 21 05:51:51 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Mar 21 05:51:51 crc kubenswrapper[4775]: > Mar 21 05:51:53 crc kubenswrapper[4775]: I0321 05:51:53.661801 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:51:53 crc kubenswrapper[4775]: E0321 05:51:53.662535 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:52:00 crc kubenswrapper[4775]: I0321 05:52:00.150342 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-snsfc" podStartSLOduration=32.012277472 podStartE2EDuration="40.150318753s" podCreationTimestamp="2026-03-21 05:51:20 +0000 UTC" firstStartedPulling="2026-03-21 05:51:22.39574965 +0000 UTC m=+3835.372213274" lastFinishedPulling="2026-03-21 05:51:30.533790931 +0000 UTC m=+3843.510254555" observedRunningTime="2026-03-21 05:51:31.55570931 +0000 UTC m=+3844.532172944" watchObservedRunningTime="2026-03-21 05:52:00.150318753 +0000 UTC m=+3873.126782397" Mar 21 05:52:00 crc kubenswrapper[4775]: I0321 05:52:00.151836 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567872-ctnt2"] Mar 21 05:52:00 crc kubenswrapper[4775]: I0321 05:52:00.153433 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567872-ctnt2" Mar 21 05:52:00 crc kubenswrapper[4775]: I0321 05:52:00.155850 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:52:00 crc kubenswrapper[4775]: I0321 05:52:00.156146 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:52:00 crc kubenswrapper[4775]: I0321 05:52:00.158829 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:52:00 crc kubenswrapper[4775]: I0321 05:52:00.163897 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567872-ctnt2"] Mar 21 05:52:00 crc kubenswrapper[4775]: I0321 05:52:00.201883 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5npp\" (UniqueName: \"kubernetes.io/projected/b19b93e1-3fea-4999-8e31-ba5cc64c91ea-kube-api-access-v5npp\") pod \"auto-csr-approver-29567872-ctnt2\" (UID: \"b19b93e1-3fea-4999-8e31-ba5cc64c91ea\") " pod="openshift-infra/auto-csr-approver-29567872-ctnt2" Mar 21 05:52:00 crc kubenswrapper[4775]: I0321 05:52:00.303271 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5npp\" (UniqueName: \"kubernetes.io/projected/b19b93e1-3fea-4999-8e31-ba5cc64c91ea-kube-api-access-v5npp\") pod \"auto-csr-approver-29567872-ctnt2\" (UID: \"b19b93e1-3fea-4999-8e31-ba5cc64c91ea\") " pod="openshift-infra/auto-csr-approver-29567872-ctnt2" Mar 21 05:52:00 crc kubenswrapper[4775]: I0321 05:52:00.329329 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5npp\" (UniqueName: \"kubernetes.io/projected/b19b93e1-3fea-4999-8e31-ba5cc64c91ea-kube-api-access-v5npp\") pod \"auto-csr-approver-29567872-ctnt2\" (UID: \"b19b93e1-3fea-4999-8e31-ba5cc64c91ea\") " pod="openshift-infra/auto-csr-approver-29567872-ctnt2" Mar 21 05:52:00 crc kubenswrapper[4775]: I0321 05:52:00.471331 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567872-ctnt2" Mar 21 05:52:00 crc kubenswrapper[4775]: I0321 05:52:00.933258 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567872-ctnt2"] Mar 21 05:52:01 crc kubenswrapper[4775]: I0321 05:52:01.812557 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567872-ctnt2" event={"ID":"b19b93e1-3fea-4999-8e31-ba5cc64c91ea","Type":"ContainerStarted","Data":"f7d5d467cc0169abf2339a995d4bae6f7dc8fca7c98d6a4d83edb19d0351f20b"} Mar 21 05:52:01 crc kubenswrapper[4775]: I0321 05:52:01.826469 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-snsfc" podUID="79dd9ac3-231b-4217-a717-a3d127b0c0ca" containerName="registry-server" probeResult="failure" output=< Mar 21 05:52:01 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Mar 21 05:52:01 crc kubenswrapper[4775]: > Mar 21 05:52:03 crc kubenswrapper[4775]: I0321 05:52:03.839061 4775 generic.go:334] "Generic (PLEG): container finished" podID="b19b93e1-3fea-4999-8e31-ba5cc64c91ea" containerID="25394576ff3d5198021128077858159917ead74d0d3b8d2389ae8fb099e54be8" exitCode=0 Mar 21 05:52:03 crc kubenswrapper[4775]: I0321 05:52:03.839439 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567872-ctnt2" event={"ID":"b19b93e1-3fea-4999-8e31-ba5cc64c91ea","Type":"ContainerDied","Data":"25394576ff3d5198021128077858159917ead74d0d3b8d2389ae8fb099e54be8"} Mar 21 05:52:05 crc kubenswrapper[4775]: I0321 05:52:05.233849 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567872-ctnt2" Mar 21 05:52:05 crc kubenswrapper[4775]: I0321 05:52:05.424921 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5npp\" (UniqueName: \"kubernetes.io/projected/b19b93e1-3fea-4999-8e31-ba5cc64c91ea-kube-api-access-v5npp\") pod \"b19b93e1-3fea-4999-8e31-ba5cc64c91ea\" (UID: \"b19b93e1-3fea-4999-8e31-ba5cc64c91ea\") " Mar 21 05:52:05 crc kubenswrapper[4775]: I0321 05:52:05.431767 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b19b93e1-3fea-4999-8e31-ba5cc64c91ea-kube-api-access-v5npp" (OuterVolumeSpecName: "kube-api-access-v5npp") pod "b19b93e1-3fea-4999-8e31-ba5cc64c91ea" (UID: "b19b93e1-3fea-4999-8e31-ba5cc64c91ea"). InnerVolumeSpecName "kube-api-access-v5npp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:52:05 crc kubenswrapper[4775]: I0321 05:52:05.527871 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5npp\" (UniqueName: \"kubernetes.io/projected/b19b93e1-3fea-4999-8e31-ba5cc64c91ea-kube-api-access-v5npp\") on node \"crc\" DevicePath \"\"" Mar 21 05:52:05 crc kubenswrapper[4775]: I0321 05:52:05.661751 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:52:05 crc kubenswrapper[4775]: E0321 05:52:05.662104 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:52:05 crc kubenswrapper[4775]: I0321 05:52:05.865069 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567872-ctnt2" event={"ID":"b19b93e1-3fea-4999-8e31-ba5cc64c91ea","Type":"ContainerDied","Data":"f7d5d467cc0169abf2339a995d4bae6f7dc8fca7c98d6a4d83edb19d0351f20b"} Mar 21 05:52:05 crc kubenswrapper[4775]: I0321 05:52:05.865160 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7d5d467cc0169abf2339a995d4bae6f7dc8fca7c98d6a4d83edb19d0351f20b" Mar 21 05:52:05 crc kubenswrapper[4775]: I0321 05:52:05.865180 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567872-ctnt2" Mar 21 05:52:06 crc kubenswrapper[4775]: I0321 05:52:06.322450 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567866-r9qrw"] Mar 21 05:52:06 crc kubenswrapper[4775]: I0321 05:52:06.331086 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567866-r9qrw"] Mar 21 05:52:07 crc kubenswrapper[4775]: I0321 05:52:07.676983 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52968d61-f2a8-4bcd-9e97-cb01c508a52e" path="/var/lib/kubelet/pods/52968d61-f2a8-4bcd-9e97-cb01c508a52e/volumes" Mar 21 05:52:11 crc kubenswrapper[4775]: I0321 05:52:11.867427 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-snsfc" podUID="79dd9ac3-231b-4217-a717-a3d127b0c0ca" containerName="registry-server" probeResult="failure" output=< Mar 21 05:52:11 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Mar 21 05:52:11 crc kubenswrapper[4775]: > Mar 21 05:52:18 crc kubenswrapper[4775]: I0321 05:52:18.661666 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:52:18 crc kubenswrapper[4775]: E0321 05:52:18.662392 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:52:21 crc kubenswrapper[4775]: I0321 05:52:21.838900 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-snsfc" podUID="79dd9ac3-231b-4217-a717-a3d127b0c0ca" containerName="registry-server" probeResult="failure" output=< Mar 21 05:52:21 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Mar 21 05:52:21 crc kubenswrapper[4775]: > Mar 21 05:52:31 crc kubenswrapper[4775]: I0321 05:52:31.836355 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-snsfc" podUID="79dd9ac3-231b-4217-a717-a3d127b0c0ca" containerName="registry-server" probeResult="failure" output=< Mar 21 05:52:31 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Mar 21 05:52:31 crc kubenswrapper[4775]: > Mar 21 05:52:33 crc kubenswrapper[4775]: I0321 05:52:33.661330 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:52:33 crc kubenswrapper[4775]: E0321 05:52:33.662030 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:52:40 crc kubenswrapper[4775]: I0321 05:52:40.836305 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-snsfc" Mar 21 05:52:40 crc kubenswrapper[4775]: I0321 05:52:40.896629 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-snsfc" Mar 21 05:52:41 crc kubenswrapper[4775]: I0321 05:52:41.074939 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-snsfc"] Mar 21 05:52:42 crc kubenswrapper[4775]: I0321 05:52:42.196358 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-snsfc" podUID="79dd9ac3-231b-4217-a717-a3d127b0c0ca" containerName="registry-server" containerID="cri-o://a9c141457cd1a8a69f1b8a93fbbbea5921c7cfed496f1b33f8197ea1b1d27341" gracePeriod=2 Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.156334 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-snsfc" Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.208360 4775 generic.go:334] "Generic (PLEG): container finished" podID="79dd9ac3-231b-4217-a717-a3d127b0c0ca" containerID="a9c141457cd1a8a69f1b8a93fbbbea5921c7cfed496f1b33f8197ea1b1d27341" exitCode=0 Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.208406 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snsfc" event={"ID":"79dd9ac3-231b-4217-a717-a3d127b0c0ca","Type":"ContainerDied","Data":"a9c141457cd1a8a69f1b8a93fbbbea5921c7cfed496f1b33f8197ea1b1d27341"} Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.208435 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snsfc" event={"ID":"79dd9ac3-231b-4217-a717-a3d127b0c0ca","Type":"ContainerDied","Data":"60650042fc6959477421db51b293d07bc331592b53cc199339128f46e38e52de"} Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.208453 4775 scope.go:117] "RemoveContainer" containerID="a9c141457cd1a8a69f1b8a93fbbbea5921c7cfed496f1b33f8197ea1b1d27341" Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.208604 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-snsfc" Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.234779 4775 scope.go:117] "RemoveContainer" containerID="6b761fc8b5585d54f920ab7f638b663e14b509ba0f6869c22d0a1c572cfc5a8d" Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.242443 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79dd9ac3-231b-4217-a717-a3d127b0c0ca-utilities\") pod \"79dd9ac3-231b-4217-a717-a3d127b0c0ca\" (UID: \"79dd9ac3-231b-4217-a717-a3d127b0c0ca\") " Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.242629 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqhpz\" (UniqueName: \"kubernetes.io/projected/79dd9ac3-231b-4217-a717-a3d127b0c0ca-kube-api-access-jqhpz\") pod \"79dd9ac3-231b-4217-a717-a3d127b0c0ca\" (UID: \"79dd9ac3-231b-4217-a717-a3d127b0c0ca\") " Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.242745 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79dd9ac3-231b-4217-a717-a3d127b0c0ca-catalog-content\") pod \"79dd9ac3-231b-4217-a717-a3d127b0c0ca\" (UID: \"79dd9ac3-231b-4217-a717-a3d127b0c0ca\") " Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.243457 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79dd9ac3-231b-4217-a717-a3d127b0c0ca-utilities" (OuterVolumeSpecName: "utilities") pod "79dd9ac3-231b-4217-a717-a3d127b0c0ca" (UID: "79dd9ac3-231b-4217-a717-a3d127b0c0ca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.250797 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79dd9ac3-231b-4217-a717-a3d127b0c0ca-kube-api-access-jqhpz" (OuterVolumeSpecName: "kube-api-access-jqhpz") pod "79dd9ac3-231b-4217-a717-a3d127b0c0ca" (UID: "79dd9ac3-231b-4217-a717-a3d127b0c0ca"). InnerVolumeSpecName "kube-api-access-jqhpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.259313 4775 scope.go:117] "RemoveContainer" containerID="a695039e53bef0f3edbd44ff68e271754c9a8e6448d909abcaca31bcc1cb9ed6" Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.345835 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqhpz\" (UniqueName: \"kubernetes.io/projected/79dd9ac3-231b-4217-a717-a3d127b0c0ca-kube-api-access-jqhpz\") on node \"crc\" DevicePath \"\"" Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.345903 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79dd9ac3-231b-4217-a717-a3d127b0c0ca-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.354475 4775 scope.go:117] "RemoveContainer" containerID="a9c141457cd1a8a69f1b8a93fbbbea5921c7cfed496f1b33f8197ea1b1d27341" Mar 21 05:52:43 crc kubenswrapper[4775]: E0321 05:52:43.354901 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9c141457cd1a8a69f1b8a93fbbbea5921c7cfed496f1b33f8197ea1b1d27341\": container with ID starting with a9c141457cd1a8a69f1b8a93fbbbea5921c7cfed496f1b33f8197ea1b1d27341 not found: ID does not exist" containerID="a9c141457cd1a8a69f1b8a93fbbbea5921c7cfed496f1b33f8197ea1b1d27341" Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.354946 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9c141457cd1a8a69f1b8a93fbbbea5921c7cfed496f1b33f8197ea1b1d27341"} err="failed to get container status \"a9c141457cd1a8a69f1b8a93fbbbea5921c7cfed496f1b33f8197ea1b1d27341\": rpc error: code = NotFound desc = could not find container \"a9c141457cd1a8a69f1b8a93fbbbea5921c7cfed496f1b33f8197ea1b1d27341\": container with ID starting with a9c141457cd1a8a69f1b8a93fbbbea5921c7cfed496f1b33f8197ea1b1d27341 not found: ID does not exist" Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.354975 4775 scope.go:117] "RemoveContainer" containerID="6b761fc8b5585d54f920ab7f638b663e14b509ba0f6869c22d0a1c572cfc5a8d" Mar 21 05:52:43 crc kubenswrapper[4775]: E0321 05:52:43.355282 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b761fc8b5585d54f920ab7f638b663e14b509ba0f6869c22d0a1c572cfc5a8d\": container with ID starting with 6b761fc8b5585d54f920ab7f638b663e14b509ba0f6869c22d0a1c572cfc5a8d not found: ID does not exist" containerID="6b761fc8b5585d54f920ab7f638b663e14b509ba0f6869c22d0a1c572cfc5a8d" Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.355333 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b761fc8b5585d54f920ab7f638b663e14b509ba0f6869c22d0a1c572cfc5a8d"} err="failed to get container status \"6b761fc8b5585d54f920ab7f638b663e14b509ba0f6869c22d0a1c572cfc5a8d\": rpc error: code = NotFound desc = could not find container \"6b761fc8b5585d54f920ab7f638b663e14b509ba0f6869c22d0a1c572cfc5a8d\": container with ID starting with 6b761fc8b5585d54f920ab7f638b663e14b509ba0f6869c22d0a1c572cfc5a8d not found: ID does not exist" Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.355367 4775 scope.go:117] "RemoveContainer" containerID="a695039e53bef0f3edbd44ff68e271754c9a8e6448d909abcaca31bcc1cb9ed6" Mar 21 05:52:43 crc kubenswrapper[4775]: E0321 05:52:43.355684 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a695039e53bef0f3edbd44ff68e271754c9a8e6448d909abcaca31bcc1cb9ed6\": container with ID starting with a695039e53bef0f3edbd44ff68e271754c9a8e6448d909abcaca31bcc1cb9ed6 not found: ID does not exist" containerID="a695039e53bef0f3edbd44ff68e271754c9a8e6448d909abcaca31bcc1cb9ed6" Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.355716 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a695039e53bef0f3edbd44ff68e271754c9a8e6448d909abcaca31bcc1cb9ed6"} err="failed to get container status \"a695039e53bef0f3edbd44ff68e271754c9a8e6448d909abcaca31bcc1cb9ed6\": rpc error: code = NotFound desc = could not find container \"a695039e53bef0f3edbd44ff68e271754c9a8e6448d909abcaca31bcc1cb9ed6\": container with ID starting with a695039e53bef0f3edbd44ff68e271754c9a8e6448d909abcaca31bcc1cb9ed6 not found: ID does not exist" Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.378912 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79dd9ac3-231b-4217-a717-a3d127b0c0ca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "79dd9ac3-231b-4217-a717-a3d127b0c0ca" (UID: "79dd9ac3-231b-4217-a717-a3d127b0c0ca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.448862 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79dd9ac3-231b-4217-a717-a3d127b0c0ca-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.545534 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-snsfc"] Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.554383 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-snsfc"] Mar 21 05:52:43 crc kubenswrapper[4775]: I0321 05:52:43.674334 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79dd9ac3-231b-4217-a717-a3d127b0c0ca" path="/var/lib/kubelet/pods/79dd9ac3-231b-4217-a717-a3d127b0c0ca/volumes" Mar 21 05:52:46 crc kubenswrapper[4775]: I0321 05:52:46.661717 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:52:46 crc kubenswrapper[4775]: E0321 05:52:46.662329 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:52:49 crc kubenswrapper[4775]: I0321 05:52:49.364570 4775 scope.go:117] "RemoveContainer" containerID="875a527a3f14334ab80b89530523916cfe6a1d4008ed65204aa4a0f63ca04790" Mar 21 05:53:00 crc kubenswrapper[4775]: I0321 05:53:00.662818 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:53:00 crc kubenswrapper[4775]: E0321 05:53:00.663867 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:53:12 crc kubenswrapper[4775]: I0321 05:53:12.472370 4775 generic.go:334] "Generic (PLEG): container finished" podID="77cb0b91-097f-4e81-888d-57b2d37399bd" containerID="891104e8ee294ddf739928b04e0a1b0e8f6bf51e7a90dfe50eb44a5a6ed19b18" exitCode=0 Mar 21 05:53:12 crc kubenswrapper[4775]: I0321 05:53:12.472451 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b8t2r/must-gather-znpkx" event={"ID":"77cb0b91-097f-4e81-888d-57b2d37399bd","Type":"ContainerDied","Data":"891104e8ee294ddf739928b04e0a1b0e8f6bf51e7a90dfe50eb44a5a6ed19b18"} Mar 21 05:53:12 crc kubenswrapper[4775]: I0321 05:53:12.473563 4775 scope.go:117] "RemoveContainer" containerID="891104e8ee294ddf739928b04e0a1b0e8f6bf51e7a90dfe50eb44a5a6ed19b18" Mar 21 05:53:12 crc kubenswrapper[4775]: I0321 05:53:12.661923 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:53:12 crc kubenswrapper[4775]: E0321 05:53:12.662219 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:53:13 crc kubenswrapper[4775]: I0321 05:53:13.059081 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-b8t2r_must-gather-znpkx_77cb0b91-097f-4e81-888d-57b2d37399bd/gather/0.log" Mar 21 05:53:20 crc kubenswrapper[4775]: I0321 05:53:20.944329 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-b8t2r/must-gather-znpkx"] Mar 21 05:53:20 crc kubenswrapper[4775]: I0321 05:53:20.945279 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-b8t2r/must-gather-znpkx" podUID="77cb0b91-097f-4e81-888d-57b2d37399bd" containerName="copy" containerID="cri-o://76eca6bf5897fa3d0bd946a5ac8c7c93a3f123f745216d5491ccbb010515a8cc" gracePeriod=2 Mar 21 05:53:20 crc kubenswrapper[4775]: I0321 05:53:20.965528 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-b8t2r/must-gather-znpkx"] Mar 21 05:53:21 crc kubenswrapper[4775]: I0321 05:53:21.559620 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-b8t2r_must-gather-znpkx_77cb0b91-097f-4e81-888d-57b2d37399bd/copy/0.log" Mar 21 05:53:21 crc kubenswrapper[4775]: I0321 05:53:21.560364 4775 generic.go:334] "Generic (PLEG): container finished" podID="77cb0b91-097f-4e81-888d-57b2d37399bd" containerID="76eca6bf5897fa3d0bd946a5ac8c7c93a3f123f745216d5491ccbb010515a8cc" exitCode=143 Mar 21 05:53:22 crc kubenswrapper[4775]: I0321 05:53:22.171361 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-b8t2r_must-gather-znpkx_77cb0b91-097f-4e81-888d-57b2d37399bd/copy/0.log" Mar 21 05:53:22 crc kubenswrapper[4775]: I0321 05:53:22.172070 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b8t2r/must-gather-znpkx" Mar 21 05:53:22 crc kubenswrapper[4775]: I0321 05:53:22.277782 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cd8wn\" (UniqueName: \"kubernetes.io/projected/77cb0b91-097f-4e81-888d-57b2d37399bd-kube-api-access-cd8wn\") pod \"77cb0b91-097f-4e81-888d-57b2d37399bd\" (UID: \"77cb0b91-097f-4e81-888d-57b2d37399bd\") " Mar 21 05:53:22 crc kubenswrapper[4775]: I0321 05:53:22.278141 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/77cb0b91-097f-4e81-888d-57b2d37399bd-must-gather-output\") pod \"77cb0b91-097f-4e81-888d-57b2d37399bd\" (UID: \"77cb0b91-097f-4e81-888d-57b2d37399bd\") " Mar 21 05:53:22 crc kubenswrapper[4775]: I0321 05:53:22.294567 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77cb0b91-097f-4e81-888d-57b2d37399bd-kube-api-access-cd8wn" (OuterVolumeSpecName: "kube-api-access-cd8wn") pod "77cb0b91-097f-4e81-888d-57b2d37399bd" (UID: "77cb0b91-097f-4e81-888d-57b2d37399bd"). InnerVolumeSpecName "kube-api-access-cd8wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:53:22 crc kubenswrapper[4775]: I0321 05:53:22.381075 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cd8wn\" (UniqueName: \"kubernetes.io/projected/77cb0b91-097f-4e81-888d-57b2d37399bd-kube-api-access-cd8wn\") on node \"crc\" DevicePath \"\"" Mar 21 05:53:22 crc kubenswrapper[4775]: I0321 05:53:22.446696 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77cb0b91-097f-4e81-888d-57b2d37399bd-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "77cb0b91-097f-4e81-888d-57b2d37399bd" (UID: "77cb0b91-097f-4e81-888d-57b2d37399bd"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:53:22 crc kubenswrapper[4775]: I0321 05:53:22.483595 4775 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/77cb0b91-097f-4e81-888d-57b2d37399bd-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 21 05:53:22 crc kubenswrapper[4775]: I0321 05:53:22.574096 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-b8t2r_must-gather-znpkx_77cb0b91-097f-4e81-888d-57b2d37399bd/copy/0.log" Mar 21 05:53:22 crc kubenswrapper[4775]: I0321 05:53:22.574509 4775 scope.go:117] "RemoveContainer" containerID="76eca6bf5897fa3d0bd946a5ac8c7c93a3f123f745216d5491ccbb010515a8cc" Mar 21 05:53:22 crc kubenswrapper[4775]: I0321 05:53:22.574665 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b8t2r/must-gather-znpkx" Mar 21 05:53:22 crc kubenswrapper[4775]: I0321 05:53:22.607629 4775 scope.go:117] "RemoveContainer" containerID="891104e8ee294ddf739928b04e0a1b0e8f6bf51e7a90dfe50eb44a5a6ed19b18" Mar 21 05:53:23 crc kubenswrapper[4775]: I0321 05:53:23.677998 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77cb0b91-097f-4e81-888d-57b2d37399bd" path="/var/lib/kubelet/pods/77cb0b91-097f-4e81-888d-57b2d37399bd/volumes" Mar 21 05:53:25 crc kubenswrapper[4775]: I0321 05:53:25.662334 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:53:25 crc kubenswrapper[4775]: E0321 05:53:25.662732 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:53:39 crc kubenswrapper[4775]: I0321 05:53:39.660999 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:53:39 crc kubenswrapper[4775]: E0321 05:53:39.661907 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:53:54 crc kubenswrapper[4775]: I0321 05:53:54.662409 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:53:54 crc kubenswrapper[4775]: E0321 05:53:54.663305 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 05:54:00 crc kubenswrapper[4775]: I0321 05:54:00.147788 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567874-8f686"] Mar 21 05:54:00 crc kubenswrapper[4775]: E0321 05:54:00.150474 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b19b93e1-3fea-4999-8e31-ba5cc64c91ea" containerName="oc" Mar 21 05:54:00 crc kubenswrapper[4775]: I0321 05:54:00.150510 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b19b93e1-3fea-4999-8e31-ba5cc64c91ea" containerName="oc" Mar 21 05:54:00 crc kubenswrapper[4775]: E0321 05:54:00.150529 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79dd9ac3-231b-4217-a717-a3d127b0c0ca" containerName="registry-server" Mar 21 05:54:00 crc kubenswrapper[4775]: I0321 05:54:00.150537 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="79dd9ac3-231b-4217-a717-a3d127b0c0ca" containerName="registry-server" Mar 21 05:54:00 crc kubenswrapper[4775]: E0321 05:54:00.150559 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79dd9ac3-231b-4217-a717-a3d127b0c0ca" containerName="extract-content" Mar 21 05:54:00 crc kubenswrapper[4775]: I0321 05:54:00.150565 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="79dd9ac3-231b-4217-a717-a3d127b0c0ca" containerName="extract-content" Mar 21 05:54:00 crc kubenswrapper[4775]: E0321 05:54:00.150574 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77cb0b91-097f-4e81-888d-57b2d37399bd" containerName="gather" Mar 21 05:54:00 crc kubenswrapper[4775]: I0321 05:54:00.150582 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="77cb0b91-097f-4e81-888d-57b2d37399bd" containerName="gather" Mar 21 05:54:00 crc kubenswrapper[4775]: E0321 05:54:00.150597 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77cb0b91-097f-4e81-888d-57b2d37399bd" containerName="copy" Mar 21 05:54:00 crc kubenswrapper[4775]: I0321 05:54:00.150604 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="77cb0b91-097f-4e81-888d-57b2d37399bd" containerName="copy" Mar 21 05:54:00 crc kubenswrapper[4775]: E0321 05:54:00.150626 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79dd9ac3-231b-4217-a717-a3d127b0c0ca" containerName="extract-utilities" Mar 21 05:54:00 crc kubenswrapper[4775]: I0321 05:54:00.150632 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="79dd9ac3-231b-4217-a717-a3d127b0c0ca" containerName="extract-utilities" Mar 21 05:54:00 crc kubenswrapper[4775]: I0321 05:54:00.150834 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="79dd9ac3-231b-4217-a717-a3d127b0c0ca" containerName="registry-server" Mar 21 05:54:00 crc kubenswrapper[4775]: I0321 05:54:00.150848 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="77cb0b91-097f-4e81-888d-57b2d37399bd" containerName="gather" Mar 21 05:54:00 crc kubenswrapper[4775]: I0321 05:54:00.150866 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="77cb0b91-097f-4e81-888d-57b2d37399bd" containerName="copy" Mar 21 05:54:00 crc kubenswrapper[4775]: I0321 05:54:00.150876 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b19b93e1-3fea-4999-8e31-ba5cc64c91ea" containerName="oc" Mar 21 05:54:00 crc kubenswrapper[4775]: I0321 05:54:00.151666 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567874-8f686" Mar 21 05:54:00 crc kubenswrapper[4775]: I0321 05:54:00.153481 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:54:00 crc kubenswrapper[4775]: I0321 05:54:00.154559 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:54:00 crc kubenswrapper[4775]: I0321 05:54:00.154587 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:54:00 crc kubenswrapper[4775]: I0321 05:54:00.161567 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567874-8f686"] Mar 21 05:54:00 crc kubenswrapper[4775]: I0321 05:54:00.327437 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw286\" (UniqueName: \"kubernetes.io/projected/a6ebee46-23d7-4d38-946e-10f7a3238243-kube-api-access-rw286\") pod \"auto-csr-approver-29567874-8f686\" (UID: \"a6ebee46-23d7-4d38-946e-10f7a3238243\") " pod="openshift-infra/auto-csr-approver-29567874-8f686" Mar 21 05:54:00 crc kubenswrapper[4775]: I0321 05:54:00.429296 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw286\" (UniqueName: \"kubernetes.io/projected/a6ebee46-23d7-4d38-946e-10f7a3238243-kube-api-access-rw286\") pod \"auto-csr-approver-29567874-8f686\" (UID: \"a6ebee46-23d7-4d38-946e-10f7a3238243\") " pod="openshift-infra/auto-csr-approver-29567874-8f686" Mar 21 05:54:00 crc kubenswrapper[4775]: I0321 05:54:00.451106 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw286\" (UniqueName: \"kubernetes.io/projected/a6ebee46-23d7-4d38-946e-10f7a3238243-kube-api-access-rw286\") pod \"auto-csr-approver-29567874-8f686\" (UID: \"a6ebee46-23d7-4d38-946e-10f7a3238243\") " pod="openshift-infra/auto-csr-approver-29567874-8f686" Mar 21 05:54:00 crc kubenswrapper[4775]: I0321 05:54:00.476402 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567874-8f686" Mar 21 05:54:01 crc kubenswrapper[4775]: I0321 05:54:01.091636 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567874-8f686"] Mar 21 05:54:01 crc kubenswrapper[4775]: I0321 05:54:01.961744 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567874-8f686" event={"ID":"a6ebee46-23d7-4d38-946e-10f7a3238243","Type":"ContainerStarted","Data":"82a5c3a691a3d2dd7f12928070512565332eb06c5d5ceac43233d9692df468a7"} Mar 21 05:54:02 crc kubenswrapper[4775]: E0321 05:54:02.869812 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6ebee46_23d7_4d38_946e_10f7a3238243.slice/crio-conmon-b68cd9cb06e02040464cbd016f2e9b85be0a632889c514eaecf91f324598c72a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6ebee46_23d7_4d38_946e_10f7a3238243.slice/crio-b68cd9cb06e02040464cbd016f2e9b85be0a632889c514eaecf91f324598c72a.scope\": RecentStats: unable to find data in memory cache]" Mar 21 05:54:02 crc kubenswrapper[4775]: I0321 05:54:02.972827 4775 generic.go:334] "Generic (PLEG): container finished" podID="a6ebee46-23d7-4d38-946e-10f7a3238243" containerID="b68cd9cb06e02040464cbd016f2e9b85be0a632889c514eaecf91f324598c72a" exitCode=0 Mar 21 05:54:02 crc kubenswrapper[4775]: I0321 05:54:02.972879 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567874-8f686" event={"ID":"a6ebee46-23d7-4d38-946e-10f7a3238243","Type":"ContainerDied","Data":"b68cd9cb06e02040464cbd016f2e9b85be0a632889c514eaecf91f324598c72a"} Mar 21 05:54:04 crc kubenswrapper[4775]: I0321 05:54:04.366817 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567874-8f686" Mar 21 05:54:04 crc kubenswrapper[4775]: I0321 05:54:04.521722 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw286\" (UniqueName: \"kubernetes.io/projected/a6ebee46-23d7-4d38-946e-10f7a3238243-kube-api-access-rw286\") pod \"a6ebee46-23d7-4d38-946e-10f7a3238243\" (UID: \"a6ebee46-23d7-4d38-946e-10f7a3238243\") " Mar 21 05:54:04 crc kubenswrapper[4775]: I0321 05:54:04.527979 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6ebee46-23d7-4d38-946e-10f7a3238243-kube-api-access-rw286" (OuterVolumeSpecName: "kube-api-access-rw286") pod "a6ebee46-23d7-4d38-946e-10f7a3238243" (UID: "a6ebee46-23d7-4d38-946e-10f7a3238243"). InnerVolumeSpecName "kube-api-access-rw286". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:54:04 crc kubenswrapper[4775]: I0321 05:54:04.624335 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw286\" (UniqueName: \"kubernetes.io/projected/a6ebee46-23d7-4d38-946e-10f7a3238243-kube-api-access-rw286\") on node \"crc\" DevicePath \"\"" Mar 21 05:54:04 crc kubenswrapper[4775]: I0321 05:54:04.989701 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567874-8f686" event={"ID":"a6ebee46-23d7-4d38-946e-10f7a3238243","Type":"ContainerDied","Data":"82a5c3a691a3d2dd7f12928070512565332eb06c5d5ceac43233d9692df468a7"} Mar 21 05:54:04 crc kubenswrapper[4775]: I0321 05:54:04.989739 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82a5c3a691a3d2dd7f12928070512565332eb06c5d5ceac43233d9692df468a7" Mar 21 05:54:04 crc kubenswrapper[4775]: I0321 05:54:04.989820 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567874-8f686" Mar 21 05:54:05 crc kubenswrapper[4775]: I0321 05:54:05.453762 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567868-8z86d"] Mar 21 05:54:05 crc kubenswrapper[4775]: I0321 05:54:05.462244 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567868-8z86d"] Mar 21 05:54:05 crc kubenswrapper[4775]: I0321 05:54:05.671677 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdeb3ead-3f1f-491f-9a15-f550065d18fb" path="/var/lib/kubelet/pods/bdeb3ead-3f1f-491f-9a15-f550065d18fb/volumes" Mar 21 05:54:07 crc kubenswrapper[4775]: I0321 05:54:07.670779 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:54:08 crc kubenswrapper[4775]: I0321 05:54:08.021734 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerStarted","Data":"fb650740af569c16423f24bd3abda67b78d6d33afa7ae7aa33a8a75f71e85f1c"} Mar 21 05:54:49 crc kubenswrapper[4775]: I0321 05:54:49.486974 4775 scope.go:117] "RemoveContainer" containerID="6995f01e339d899d20a0310c4264870e5b4687e152b20b7bc6f33c17d3fccaea" Mar 21 05:55:45 crc kubenswrapper[4775]: I0321 05:55:45.623755 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xwtsn"] Mar 21 05:55:45 crc kubenswrapper[4775]: E0321 05:55:45.624780 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6ebee46-23d7-4d38-946e-10f7a3238243" containerName="oc" Mar 21 05:55:45 crc kubenswrapper[4775]: I0321 05:55:45.624795 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6ebee46-23d7-4d38-946e-10f7a3238243" containerName="oc" Mar 21 05:55:45 crc kubenswrapper[4775]: I0321 05:55:45.624994 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6ebee46-23d7-4d38-946e-10f7a3238243" containerName="oc" Mar 21 05:55:45 crc kubenswrapper[4775]: I0321 05:55:45.627469 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xwtsn" Mar 21 05:55:45 crc kubenswrapper[4775]: I0321 05:55:45.643246 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xwtsn"] Mar 21 05:55:45 crc kubenswrapper[4775]: I0321 05:55:45.762677 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2cc63cf-968e-4695-bb7a-bd8938511c29-catalog-content\") pod \"certified-operators-xwtsn\" (UID: \"a2cc63cf-968e-4695-bb7a-bd8938511c29\") " pod="openshift-marketplace/certified-operators-xwtsn" Mar 21 05:55:45 crc kubenswrapper[4775]: I0321 05:55:45.762742 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn4wz\" (UniqueName: \"kubernetes.io/projected/a2cc63cf-968e-4695-bb7a-bd8938511c29-kube-api-access-qn4wz\") pod \"certified-operators-xwtsn\" (UID: \"a2cc63cf-968e-4695-bb7a-bd8938511c29\") " pod="openshift-marketplace/certified-operators-xwtsn" Mar 21 05:55:45 crc kubenswrapper[4775]: I0321 05:55:45.762777 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2cc63cf-968e-4695-bb7a-bd8938511c29-utilities\") pod \"certified-operators-xwtsn\" (UID: \"a2cc63cf-968e-4695-bb7a-bd8938511c29\") " pod="openshift-marketplace/certified-operators-xwtsn" Mar 21 05:55:45 crc kubenswrapper[4775]: I0321 05:55:45.864673 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2cc63cf-968e-4695-bb7a-bd8938511c29-catalog-content\") pod \"certified-operators-xwtsn\" (UID: \"a2cc63cf-968e-4695-bb7a-bd8938511c29\") " pod="openshift-marketplace/certified-operators-xwtsn" Mar 21 05:55:45 crc kubenswrapper[4775]: I0321 05:55:45.864744 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn4wz\" (UniqueName: \"kubernetes.io/projected/a2cc63cf-968e-4695-bb7a-bd8938511c29-kube-api-access-qn4wz\") pod \"certified-operators-xwtsn\" (UID: \"a2cc63cf-968e-4695-bb7a-bd8938511c29\") " pod="openshift-marketplace/certified-operators-xwtsn" Mar 21 05:55:45 crc kubenswrapper[4775]: I0321 05:55:45.864782 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2cc63cf-968e-4695-bb7a-bd8938511c29-utilities\") pod \"certified-operators-xwtsn\" (UID: \"a2cc63cf-968e-4695-bb7a-bd8938511c29\") " pod="openshift-marketplace/certified-operators-xwtsn" Mar 21 05:55:45 crc kubenswrapper[4775]: I0321 05:55:45.865248 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2cc63cf-968e-4695-bb7a-bd8938511c29-catalog-content\") pod \"certified-operators-xwtsn\" (UID: \"a2cc63cf-968e-4695-bb7a-bd8938511c29\") " pod="openshift-marketplace/certified-operators-xwtsn" Mar 21 05:55:45 crc kubenswrapper[4775]: I0321 05:55:45.865481 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2cc63cf-968e-4695-bb7a-bd8938511c29-utilities\") pod \"certified-operators-xwtsn\" (UID: \"a2cc63cf-968e-4695-bb7a-bd8938511c29\") " pod="openshift-marketplace/certified-operators-xwtsn" Mar 21 05:55:45 crc kubenswrapper[4775]: I0321 05:55:45.897155 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn4wz\" (UniqueName: \"kubernetes.io/projected/a2cc63cf-968e-4695-bb7a-bd8938511c29-kube-api-access-qn4wz\") pod \"certified-operators-xwtsn\" (UID: \"a2cc63cf-968e-4695-bb7a-bd8938511c29\") " pod="openshift-marketplace/certified-operators-xwtsn" Mar 21 05:55:45 crc kubenswrapper[4775]: I0321 05:55:45.958783 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xwtsn" Mar 21 05:55:46 crc kubenswrapper[4775]: I0321 05:55:46.459806 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xwtsn"] Mar 21 05:55:46 crc kubenswrapper[4775]: I0321 05:55:46.971949 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwtsn" event={"ID":"a2cc63cf-968e-4695-bb7a-bd8938511c29","Type":"ContainerStarted","Data":"e43955e05be1c11e78ad56fed49e1899127312a909955f2f2464f37be4e186fd"} Mar 21 05:55:46 crc kubenswrapper[4775]: I0321 05:55:46.972286 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwtsn" event={"ID":"a2cc63cf-968e-4695-bb7a-bd8938511c29","Type":"ContainerStarted","Data":"61e7a835618349eda7c91ee47d139dfad4d5f13ad9f24f1aa5d97237bd5616ae"} Mar 21 05:55:47 crc kubenswrapper[4775]: I0321 05:55:47.980791 4775 generic.go:334] "Generic (PLEG): container finished" podID="a2cc63cf-968e-4695-bb7a-bd8938511c29" containerID="e43955e05be1c11e78ad56fed49e1899127312a909955f2f2464f37be4e186fd" exitCode=0 Mar 21 05:55:47 crc kubenswrapper[4775]: I0321 05:55:47.981142 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwtsn" event={"ID":"a2cc63cf-968e-4695-bb7a-bd8938511c29","Type":"ContainerDied","Data":"e43955e05be1c11e78ad56fed49e1899127312a909955f2f2464f37be4e186fd"} Mar 21 05:55:47 crc kubenswrapper[4775]: I0321 05:55:47.984170 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 05:55:48 crc kubenswrapper[4775]: I0321 05:55:48.997797 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwtsn" event={"ID":"a2cc63cf-968e-4695-bb7a-bd8938511c29","Type":"ContainerStarted","Data":"3797b6886ae493c02ef28cc86da9e4074a1002593383e3ae858d4f7bf02221e4"} Mar 21 05:55:50 crc kubenswrapper[4775]: I0321 05:55:50.007382 4775 generic.go:334] "Generic (PLEG): container finished" podID="a2cc63cf-968e-4695-bb7a-bd8938511c29" containerID="3797b6886ae493c02ef28cc86da9e4074a1002593383e3ae858d4f7bf02221e4" exitCode=0 Mar 21 05:55:50 crc kubenswrapper[4775]: I0321 05:55:50.007433 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwtsn" event={"ID":"a2cc63cf-968e-4695-bb7a-bd8938511c29","Type":"ContainerDied","Data":"3797b6886ae493c02ef28cc86da9e4074a1002593383e3ae858d4f7bf02221e4"} Mar 21 05:55:51 crc kubenswrapper[4775]: I0321 05:55:51.020846 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwtsn" event={"ID":"a2cc63cf-968e-4695-bb7a-bd8938511c29","Type":"ContainerStarted","Data":"cb36d15d10b6e43c11f3066b543cd0135060dab87e1c9fb8292210fb109dc69d"} Mar 21 05:55:55 crc kubenswrapper[4775]: I0321 05:55:55.960019 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xwtsn" Mar 21 05:55:55 crc kubenswrapper[4775]: I0321 05:55:55.961617 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xwtsn" Mar 21 05:55:56 crc kubenswrapper[4775]: I0321 05:55:56.010324 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xwtsn" Mar 21 05:55:56 crc kubenswrapper[4775]: I0321 05:55:56.027025 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xwtsn" podStartSLOduration=8.391415355 podStartE2EDuration="11.027005997s" podCreationTimestamp="2026-03-21 05:55:45 +0000 UTC" firstStartedPulling="2026-03-21 05:55:47.983893001 +0000 UTC m=+4100.960356625" lastFinishedPulling="2026-03-21 05:55:50.619483643 +0000 UTC m=+4103.595947267" observedRunningTime="2026-03-21 05:55:51.046813658 +0000 UTC m=+4104.023277322" watchObservedRunningTime="2026-03-21 05:55:56.027005997 +0000 UTC m=+4109.003469621" Mar 21 05:55:56 crc kubenswrapper[4775]: I0321 05:55:56.108489 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xwtsn" Mar 21 05:55:56 crc kubenswrapper[4775]: I0321 05:55:56.249818 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xwtsn"] Mar 21 05:55:58 crc kubenswrapper[4775]: I0321 05:55:58.080203 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xwtsn" podUID="a2cc63cf-968e-4695-bb7a-bd8938511c29" containerName="registry-server" containerID="cri-o://cb36d15d10b6e43c11f3066b543cd0135060dab87e1c9fb8292210fb109dc69d" gracePeriod=2 Mar 21 05:55:58 crc kubenswrapper[4775]: I0321 05:55:58.806298 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xwtsn" Mar 21 05:55:58 crc kubenswrapper[4775]: I0321 05:55:58.923196 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2cc63cf-968e-4695-bb7a-bd8938511c29-utilities\") pod \"a2cc63cf-968e-4695-bb7a-bd8938511c29\" (UID: \"a2cc63cf-968e-4695-bb7a-bd8938511c29\") " Mar 21 05:55:58 crc kubenswrapper[4775]: I0321 05:55:58.923268 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2cc63cf-968e-4695-bb7a-bd8938511c29-catalog-content\") pod \"a2cc63cf-968e-4695-bb7a-bd8938511c29\" (UID: \"a2cc63cf-968e-4695-bb7a-bd8938511c29\") " Mar 21 05:55:58 crc kubenswrapper[4775]: I0321 05:55:58.923372 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qn4wz\" (UniqueName: \"kubernetes.io/projected/a2cc63cf-968e-4695-bb7a-bd8938511c29-kube-api-access-qn4wz\") pod \"a2cc63cf-968e-4695-bb7a-bd8938511c29\" (UID: \"a2cc63cf-968e-4695-bb7a-bd8938511c29\") " Mar 21 05:55:58 crc kubenswrapper[4775]: I0321 05:55:58.924374 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2cc63cf-968e-4695-bb7a-bd8938511c29-utilities" (OuterVolumeSpecName: "utilities") pod "a2cc63cf-968e-4695-bb7a-bd8938511c29" (UID: "a2cc63cf-968e-4695-bb7a-bd8938511c29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:55:58 crc kubenswrapper[4775]: I0321 05:55:58.929594 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2cc63cf-968e-4695-bb7a-bd8938511c29-kube-api-access-qn4wz" (OuterVolumeSpecName: "kube-api-access-qn4wz") pod "a2cc63cf-968e-4695-bb7a-bd8938511c29" (UID: "a2cc63cf-968e-4695-bb7a-bd8938511c29"). InnerVolumeSpecName "kube-api-access-qn4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:55:58 crc kubenswrapper[4775]: I0321 05:55:58.991663 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2cc63cf-968e-4695-bb7a-bd8938511c29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a2cc63cf-968e-4695-bb7a-bd8938511c29" (UID: "a2cc63cf-968e-4695-bb7a-bd8938511c29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:55:59 crc kubenswrapper[4775]: I0321 05:55:59.025735 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2cc63cf-968e-4695-bb7a-bd8938511c29-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:55:59 crc kubenswrapper[4775]: I0321 05:55:59.025790 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2cc63cf-968e-4695-bb7a-bd8938511c29-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:55:59 crc kubenswrapper[4775]: I0321 05:55:59.025808 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qn4wz\" (UniqueName: \"kubernetes.io/projected/a2cc63cf-968e-4695-bb7a-bd8938511c29-kube-api-access-qn4wz\") on node \"crc\" DevicePath \"\"" Mar 21 05:55:59 crc kubenswrapper[4775]: I0321 05:55:59.091711 4775 generic.go:334] "Generic (PLEG): container finished" podID="a2cc63cf-968e-4695-bb7a-bd8938511c29" containerID="cb36d15d10b6e43c11f3066b543cd0135060dab87e1c9fb8292210fb109dc69d" exitCode=0 Mar 21 05:55:59 crc kubenswrapper[4775]: I0321 05:55:59.091769 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwtsn" event={"ID":"a2cc63cf-968e-4695-bb7a-bd8938511c29","Type":"ContainerDied","Data":"cb36d15d10b6e43c11f3066b543cd0135060dab87e1c9fb8292210fb109dc69d"} Mar 21 05:55:59 crc kubenswrapper[4775]: I0321 05:55:59.091821 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xwtsn" Mar 21 05:55:59 crc kubenswrapper[4775]: I0321 05:55:59.091833 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwtsn" event={"ID":"a2cc63cf-968e-4695-bb7a-bd8938511c29","Type":"ContainerDied","Data":"61e7a835618349eda7c91ee47d139dfad4d5f13ad9f24f1aa5d97237bd5616ae"} Mar 21 05:55:59 crc kubenswrapper[4775]: I0321 05:55:59.091860 4775 scope.go:117] "RemoveContainer" containerID="cb36d15d10b6e43c11f3066b543cd0135060dab87e1c9fb8292210fb109dc69d" Mar 21 05:55:59 crc kubenswrapper[4775]: I0321 05:55:59.114464 4775 scope.go:117] "RemoveContainer" containerID="3797b6886ae493c02ef28cc86da9e4074a1002593383e3ae858d4f7bf02221e4" Mar 21 05:55:59 crc kubenswrapper[4775]: I0321 05:55:59.128798 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xwtsn"] Mar 21 05:55:59 crc kubenswrapper[4775]: I0321 05:55:59.138675 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xwtsn"] Mar 21 05:55:59 crc kubenswrapper[4775]: I0321 05:55:59.145317 4775 scope.go:117] "RemoveContainer" containerID="e43955e05be1c11e78ad56fed49e1899127312a909955f2f2464f37be4e186fd" Mar 21 05:55:59 crc kubenswrapper[4775]: I0321 05:55:59.186602 4775 scope.go:117] "RemoveContainer" containerID="cb36d15d10b6e43c11f3066b543cd0135060dab87e1c9fb8292210fb109dc69d" Mar 21 05:55:59 crc kubenswrapper[4775]: E0321 05:55:59.187294 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb36d15d10b6e43c11f3066b543cd0135060dab87e1c9fb8292210fb109dc69d\": container with ID starting with cb36d15d10b6e43c11f3066b543cd0135060dab87e1c9fb8292210fb109dc69d not found: ID does not exist" containerID="cb36d15d10b6e43c11f3066b543cd0135060dab87e1c9fb8292210fb109dc69d" Mar 21 05:55:59 crc kubenswrapper[4775]: I0321 05:55:59.187334 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb36d15d10b6e43c11f3066b543cd0135060dab87e1c9fb8292210fb109dc69d"} err="failed to get container status \"cb36d15d10b6e43c11f3066b543cd0135060dab87e1c9fb8292210fb109dc69d\": rpc error: code = NotFound desc = could not find container \"cb36d15d10b6e43c11f3066b543cd0135060dab87e1c9fb8292210fb109dc69d\": container with ID starting with cb36d15d10b6e43c11f3066b543cd0135060dab87e1c9fb8292210fb109dc69d not found: ID does not exist" Mar 21 05:55:59 crc kubenswrapper[4775]: I0321 05:55:59.187360 4775 scope.go:117] "RemoveContainer" containerID="3797b6886ae493c02ef28cc86da9e4074a1002593383e3ae858d4f7bf02221e4" Mar 21 05:55:59 crc kubenswrapper[4775]: E0321 05:55:59.187799 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3797b6886ae493c02ef28cc86da9e4074a1002593383e3ae858d4f7bf02221e4\": container with ID starting with 3797b6886ae493c02ef28cc86da9e4074a1002593383e3ae858d4f7bf02221e4 not found: ID does not exist" containerID="3797b6886ae493c02ef28cc86da9e4074a1002593383e3ae858d4f7bf02221e4" Mar 21 05:55:59 crc kubenswrapper[4775]: I0321 05:55:59.187826 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3797b6886ae493c02ef28cc86da9e4074a1002593383e3ae858d4f7bf02221e4"} err="failed to get container status \"3797b6886ae493c02ef28cc86da9e4074a1002593383e3ae858d4f7bf02221e4\": rpc error: code = NotFound desc = could not find container \"3797b6886ae493c02ef28cc86da9e4074a1002593383e3ae858d4f7bf02221e4\": container with ID starting with 3797b6886ae493c02ef28cc86da9e4074a1002593383e3ae858d4f7bf02221e4 not found: ID does not exist" Mar 21 05:55:59 crc kubenswrapper[4775]: I0321 05:55:59.187841 4775 scope.go:117] "RemoveContainer" containerID="e43955e05be1c11e78ad56fed49e1899127312a909955f2f2464f37be4e186fd" Mar 21 05:55:59 crc kubenswrapper[4775]: E0321 05:55:59.188285 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e43955e05be1c11e78ad56fed49e1899127312a909955f2f2464f37be4e186fd\": container with ID starting with e43955e05be1c11e78ad56fed49e1899127312a909955f2f2464f37be4e186fd not found: ID does not exist" containerID="e43955e05be1c11e78ad56fed49e1899127312a909955f2f2464f37be4e186fd" Mar 21 05:55:59 crc kubenswrapper[4775]: I0321 05:55:59.188308 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e43955e05be1c11e78ad56fed49e1899127312a909955f2f2464f37be4e186fd"} err="failed to get container status \"e43955e05be1c11e78ad56fed49e1899127312a909955f2f2464f37be4e186fd\": rpc error: code = NotFound desc = could not find container \"e43955e05be1c11e78ad56fed49e1899127312a909955f2f2464f37be4e186fd\": container with ID starting with e43955e05be1c11e78ad56fed49e1899127312a909955f2f2464f37be4e186fd not found: ID does not exist" Mar 21 05:55:59 crc kubenswrapper[4775]: I0321 05:55:59.675079 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2cc63cf-968e-4695-bb7a-bd8938511c29" path="/var/lib/kubelet/pods/a2cc63cf-968e-4695-bb7a-bd8938511c29/volumes" Mar 21 05:56:00 crc kubenswrapper[4775]: I0321 05:56:00.148437 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567876-cl4xc"] Mar 21 05:56:00 crc kubenswrapper[4775]: E0321 05:56:00.149288 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2cc63cf-968e-4695-bb7a-bd8938511c29" containerName="extract-content" Mar 21 05:56:00 crc kubenswrapper[4775]: I0321 05:56:00.149304 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2cc63cf-968e-4695-bb7a-bd8938511c29" containerName="extract-content" Mar 21 05:56:00 crc kubenswrapper[4775]: E0321 05:56:00.149349 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2cc63cf-968e-4695-bb7a-bd8938511c29" containerName="registry-server" Mar 21 05:56:00 crc kubenswrapper[4775]: I0321 05:56:00.149357 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2cc63cf-968e-4695-bb7a-bd8938511c29" containerName="registry-server" Mar 21 05:56:00 crc kubenswrapper[4775]: E0321 05:56:00.149383 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2cc63cf-968e-4695-bb7a-bd8938511c29" containerName="extract-utilities" Mar 21 05:56:00 crc kubenswrapper[4775]: I0321 05:56:00.149394 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2cc63cf-968e-4695-bb7a-bd8938511c29" containerName="extract-utilities" Mar 21 05:56:00 crc kubenswrapper[4775]: I0321 05:56:00.149686 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2cc63cf-968e-4695-bb7a-bd8938511c29" containerName="registry-server" Mar 21 05:56:00 crc kubenswrapper[4775]: I0321 05:56:00.150510 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567876-cl4xc" Mar 21 05:56:00 crc kubenswrapper[4775]: I0321 05:56:00.153187 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:56:00 crc kubenswrapper[4775]: I0321 05:56:00.153640 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:56:00 crc kubenswrapper[4775]: I0321 05:56:00.153805 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:56:00 crc kubenswrapper[4775]: I0321 05:56:00.159154 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567876-cl4xc"] Mar 21 05:56:00 crc kubenswrapper[4775]: I0321 05:56:00.251197 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw5bd\" (UniqueName: \"kubernetes.io/projected/77bcfe47-3b2b-4864-8659-69053e25c0a7-kube-api-access-rw5bd\") pod \"auto-csr-approver-29567876-cl4xc\" (UID: \"77bcfe47-3b2b-4864-8659-69053e25c0a7\") " pod="openshift-infra/auto-csr-approver-29567876-cl4xc" Mar 21 05:56:00 crc kubenswrapper[4775]: I0321 05:56:00.353589 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw5bd\" (UniqueName: \"kubernetes.io/projected/77bcfe47-3b2b-4864-8659-69053e25c0a7-kube-api-access-rw5bd\") pod \"auto-csr-approver-29567876-cl4xc\" (UID: \"77bcfe47-3b2b-4864-8659-69053e25c0a7\") " pod="openshift-infra/auto-csr-approver-29567876-cl4xc" Mar 21 05:56:00 crc kubenswrapper[4775]: I0321 05:56:00.600722 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw5bd\" (UniqueName: \"kubernetes.io/projected/77bcfe47-3b2b-4864-8659-69053e25c0a7-kube-api-access-rw5bd\") pod \"auto-csr-approver-29567876-cl4xc\" (UID: \"77bcfe47-3b2b-4864-8659-69053e25c0a7\") " pod="openshift-infra/auto-csr-approver-29567876-cl4xc" Mar 21 05:56:00 crc kubenswrapper[4775]: I0321 05:56:00.770138 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567876-cl4xc" Mar 21 05:56:01 crc kubenswrapper[4775]: W0321 05:56:01.236251 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77bcfe47_3b2b_4864_8659_69053e25c0a7.slice/crio-42715204dec9b60bfad42994e43ccd1afd991606987f22c43b1f2b84ef7f590a WatchSource:0}: Error finding container 42715204dec9b60bfad42994e43ccd1afd991606987f22c43b1f2b84ef7f590a: Status 404 returned error can't find the container with id 42715204dec9b60bfad42994e43ccd1afd991606987f22c43b1f2b84ef7f590a Mar 21 05:56:01 crc kubenswrapper[4775]: I0321 05:56:01.239099 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567876-cl4xc"] Mar 21 05:56:02 crc kubenswrapper[4775]: I0321 05:56:02.123630 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567876-cl4xc" event={"ID":"77bcfe47-3b2b-4864-8659-69053e25c0a7","Type":"ContainerStarted","Data":"42715204dec9b60bfad42994e43ccd1afd991606987f22c43b1f2b84ef7f590a"} Mar 21 05:56:03 crc kubenswrapper[4775]: I0321 05:56:03.133950 4775 generic.go:334] "Generic (PLEG): container finished" podID="77bcfe47-3b2b-4864-8659-69053e25c0a7" containerID="ff881789564104cd73abc465716a89e9559dbe987cab602aecfdcfd1bd6865c8" exitCode=0 Mar 21 05:56:03 crc kubenswrapper[4775]: I0321 05:56:03.134054 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567876-cl4xc" event={"ID":"77bcfe47-3b2b-4864-8659-69053e25c0a7","Type":"ContainerDied","Data":"ff881789564104cd73abc465716a89e9559dbe987cab602aecfdcfd1bd6865c8"} Mar 21 05:56:04 crc kubenswrapper[4775]: I0321 05:56:04.598536 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567876-cl4xc" Mar 21 05:56:04 crc kubenswrapper[4775]: I0321 05:56:04.643733 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw5bd\" (UniqueName: \"kubernetes.io/projected/77bcfe47-3b2b-4864-8659-69053e25c0a7-kube-api-access-rw5bd\") pod \"77bcfe47-3b2b-4864-8659-69053e25c0a7\" (UID: \"77bcfe47-3b2b-4864-8659-69053e25c0a7\") " Mar 21 05:56:04 crc kubenswrapper[4775]: I0321 05:56:04.653850 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77bcfe47-3b2b-4864-8659-69053e25c0a7-kube-api-access-rw5bd" (OuterVolumeSpecName: "kube-api-access-rw5bd") pod "77bcfe47-3b2b-4864-8659-69053e25c0a7" (UID: "77bcfe47-3b2b-4864-8659-69053e25c0a7"). InnerVolumeSpecName "kube-api-access-rw5bd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:56:04 crc kubenswrapper[4775]: I0321 05:56:04.747192 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw5bd\" (UniqueName: \"kubernetes.io/projected/77bcfe47-3b2b-4864-8659-69053e25c0a7-kube-api-access-rw5bd\") on node \"crc\" DevicePath \"\"" Mar 21 05:56:05 crc kubenswrapper[4775]: I0321 05:56:05.160895 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567876-cl4xc" event={"ID":"77bcfe47-3b2b-4864-8659-69053e25c0a7","Type":"ContainerDied","Data":"42715204dec9b60bfad42994e43ccd1afd991606987f22c43b1f2b84ef7f590a"} Mar 21 05:56:05 crc kubenswrapper[4775]: I0321 05:56:05.160939 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567876-cl4xc" Mar 21 05:56:05 crc kubenswrapper[4775]: I0321 05:56:05.160946 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42715204dec9b60bfad42994e43ccd1afd991606987f22c43b1f2b84ef7f590a" Mar 21 05:56:05 crc kubenswrapper[4775]: I0321 05:56:05.675266 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567870-hdgmt"] Mar 21 05:56:05 crc kubenswrapper[4775]: I0321 05:56:05.707304 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567870-hdgmt"] Mar 21 05:56:07 crc kubenswrapper[4775]: I0321 05:56:07.672817 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b54d80b-a103-474c-8214-44ec9b595a43" path="/var/lib/kubelet/pods/3b54d80b-a103-474c-8214-44ec9b595a43/volumes" Mar 21 05:56:23 crc kubenswrapper[4775]: I0321 05:56:23.245012 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hvhrr"] Mar 21 05:56:23 crc kubenswrapper[4775]: E0321 05:56:23.246325 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77bcfe47-3b2b-4864-8659-69053e25c0a7" containerName="oc" Mar 21 05:56:23 crc kubenswrapper[4775]: I0321 05:56:23.246344 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="77bcfe47-3b2b-4864-8659-69053e25c0a7" containerName="oc" Mar 21 05:56:23 crc kubenswrapper[4775]: I0321 05:56:23.247114 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="77bcfe47-3b2b-4864-8659-69053e25c0a7" containerName="oc" Mar 21 05:56:23 crc kubenswrapper[4775]: I0321 05:56:23.248913 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hvhrr" Mar 21 05:56:23 crc kubenswrapper[4775]: I0321 05:56:23.269819 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hvhrr"] Mar 21 05:56:23 crc kubenswrapper[4775]: I0321 05:56:23.301097 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9132264-94c1-407d-87ab-d7f546907220-utilities\") pod \"community-operators-hvhrr\" (UID: \"a9132264-94c1-407d-87ab-d7f546907220\") " pod="openshift-marketplace/community-operators-hvhrr" Mar 21 05:56:23 crc kubenswrapper[4775]: I0321 05:56:23.301184 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v54rl\" (UniqueName: \"kubernetes.io/projected/a9132264-94c1-407d-87ab-d7f546907220-kube-api-access-v54rl\") pod \"community-operators-hvhrr\" (UID: \"a9132264-94c1-407d-87ab-d7f546907220\") " pod="openshift-marketplace/community-operators-hvhrr" Mar 21 05:56:23 crc kubenswrapper[4775]: I0321 05:56:23.301223 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9132264-94c1-407d-87ab-d7f546907220-catalog-content\") pod \"community-operators-hvhrr\" (UID: \"a9132264-94c1-407d-87ab-d7f546907220\") " pod="openshift-marketplace/community-operators-hvhrr" Mar 21 05:56:23 crc kubenswrapper[4775]: I0321 05:56:23.402950 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v54rl\" (UniqueName: \"kubernetes.io/projected/a9132264-94c1-407d-87ab-d7f546907220-kube-api-access-v54rl\") pod \"community-operators-hvhrr\" (UID: \"a9132264-94c1-407d-87ab-d7f546907220\") " pod="openshift-marketplace/community-operators-hvhrr" Mar 21 05:56:23 crc kubenswrapper[4775]: I0321 05:56:23.403040 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9132264-94c1-407d-87ab-d7f546907220-catalog-content\") pod \"community-operators-hvhrr\" (UID: \"a9132264-94c1-407d-87ab-d7f546907220\") " pod="openshift-marketplace/community-operators-hvhrr" Mar 21 05:56:23 crc kubenswrapper[4775]: I0321 05:56:23.403207 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9132264-94c1-407d-87ab-d7f546907220-utilities\") pod \"community-operators-hvhrr\" (UID: \"a9132264-94c1-407d-87ab-d7f546907220\") " pod="openshift-marketplace/community-operators-hvhrr" Mar 21 05:56:23 crc kubenswrapper[4775]: I0321 05:56:23.403644 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9132264-94c1-407d-87ab-d7f546907220-catalog-content\") pod \"community-operators-hvhrr\" (UID: \"a9132264-94c1-407d-87ab-d7f546907220\") " pod="openshift-marketplace/community-operators-hvhrr" Mar 21 05:56:23 crc kubenswrapper[4775]: I0321 05:56:23.403683 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9132264-94c1-407d-87ab-d7f546907220-utilities\") pod \"community-operators-hvhrr\" (UID: \"a9132264-94c1-407d-87ab-d7f546907220\") " pod="openshift-marketplace/community-operators-hvhrr" Mar 21 05:56:23 crc kubenswrapper[4775]: I0321 05:56:23.429086 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v54rl\" (UniqueName: \"kubernetes.io/projected/a9132264-94c1-407d-87ab-d7f546907220-kube-api-access-v54rl\") pod \"community-operators-hvhrr\" (UID: \"a9132264-94c1-407d-87ab-d7f546907220\") " pod="openshift-marketplace/community-operators-hvhrr" Mar 21 05:56:23 crc kubenswrapper[4775]: I0321 05:56:23.581861 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hvhrr" Mar 21 05:56:24 crc kubenswrapper[4775]: I0321 05:56:24.154461 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hvhrr"] Mar 21 05:56:24 crc kubenswrapper[4775]: I0321 05:56:24.880089 4775 generic.go:334] "Generic (PLEG): container finished" podID="a9132264-94c1-407d-87ab-d7f546907220" containerID="56ec5eaa6447799381690c7e64db1dc2298dedf7d9f2b726e9f414fe1bc074cf" exitCode=0 Mar 21 05:56:24 crc kubenswrapper[4775]: I0321 05:56:24.880209 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hvhrr" event={"ID":"a9132264-94c1-407d-87ab-d7f546907220","Type":"ContainerDied","Data":"56ec5eaa6447799381690c7e64db1dc2298dedf7d9f2b726e9f414fe1bc074cf"} Mar 21 05:56:24 crc kubenswrapper[4775]: I0321 05:56:24.880456 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hvhrr" event={"ID":"a9132264-94c1-407d-87ab-d7f546907220","Type":"ContainerStarted","Data":"2585c889d397a5cc818abfd7d487d4c812588c6d0795639c8158b4cc4e2799db"} Mar 21 05:56:25 crc kubenswrapper[4775]: I0321 05:56:25.901487 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hvhrr" event={"ID":"a9132264-94c1-407d-87ab-d7f546907220","Type":"ContainerStarted","Data":"db2adef6aa111c2105ea410f3c602299796961892b8bbcc0eaa8ea40de324259"} Mar 21 05:56:27 crc kubenswrapper[4775]: I0321 05:56:27.939621 4775 generic.go:334] "Generic (PLEG): container finished" podID="a9132264-94c1-407d-87ab-d7f546907220" containerID="db2adef6aa111c2105ea410f3c602299796961892b8bbcc0eaa8ea40de324259" exitCode=0 Mar 21 05:56:27 crc kubenswrapper[4775]: I0321 05:56:27.939669 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hvhrr" event={"ID":"a9132264-94c1-407d-87ab-d7f546907220","Type":"ContainerDied","Data":"db2adef6aa111c2105ea410f3c602299796961892b8bbcc0eaa8ea40de324259"} Mar 21 05:56:28 crc kubenswrapper[4775]: I0321 05:56:28.950026 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hvhrr" event={"ID":"a9132264-94c1-407d-87ab-d7f546907220","Type":"ContainerStarted","Data":"124837a8aeed307cbf3f0d2a7f042cf5856dac1eb3286c8f4ca6d73a2fd39f43"} Mar 21 05:56:28 crc kubenswrapper[4775]: I0321 05:56:28.976427 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hvhrr" podStartSLOduration=2.482502112 podStartE2EDuration="5.976405937s" podCreationTimestamp="2026-03-21 05:56:23 +0000 UTC" firstStartedPulling="2026-03-21 05:56:24.88177193 +0000 UTC m=+4137.858235554" lastFinishedPulling="2026-03-21 05:56:28.375675755 +0000 UTC m=+4141.352139379" observedRunningTime="2026-03-21 05:56:28.970483199 +0000 UTC m=+4141.946946823" watchObservedRunningTime="2026-03-21 05:56:28.976405937 +0000 UTC m=+4141.952869561" Mar 21 05:56:32 crc kubenswrapper[4775]: I0321 05:56:32.039210 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2gd5d/must-gather-fmnth"] Mar 21 05:56:32 crc kubenswrapper[4775]: I0321 05:56:32.041878 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2gd5d/must-gather-fmnth" Mar 21 05:56:32 crc kubenswrapper[4775]: I0321 05:56:32.043834 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-2gd5d"/"openshift-service-ca.crt" Mar 21 05:56:32 crc kubenswrapper[4775]: I0321 05:56:32.043897 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-2gd5d"/"default-dockercfg-hpcs9" Mar 21 05:56:32 crc kubenswrapper[4775]: I0321 05:56:32.045411 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-2gd5d"/"kube-root-ca.crt" Mar 21 05:56:32 crc kubenswrapper[4775]: I0321 05:56:32.050624 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2gd5d/must-gather-fmnth"] Mar 21 05:56:32 crc kubenswrapper[4775]: I0321 05:56:32.065896 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5jqm\" (UniqueName: \"kubernetes.io/projected/f8e9d6cc-a02d-4747-8f7c-f439b3227b11-kube-api-access-k5jqm\") pod \"must-gather-fmnth\" (UID: \"f8e9d6cc-a02d-4747-8f7c-f439b3227b11\") " pod="openshift-must-gather-2gd5d/must-gather-fmnth" Mar 21 05:56:32 crc kubenswrapper[4775]: I0321 05:56:32.065981 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f8e9d6cc-a02d-4747-8f7c-f439b3227b11-must-gather-output\") pod \"must-gather-fmnth\" (UID: \"f8e9d6cc-a02d-4747-8f7c-f439b3227b11\") " pod="openshift-must-gather-2gd5d/must-gather-fmnth" Mar 21 05:56:32 crc kubenswrapper[4775]: I0321 05:56:32.167381 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5jqm\" (UniqueName: \"kubernetes.io/projected/f8e9d6cc-a02d-4747-8f7c-f439b3227b11-kube-api-access-k5jqm\") pod \"must-gather-fmnth\" (UID: \"f8e9d6cc-a02d-4747-8f7c-f439b3227b11\") " pod="openshift-must-gather-2gd5d/must-gather-fmnth" Mar 21 05:56:32 crc kubenswrapper[4775]: I0321 05:56:32.167461 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f8e9d6cc-a02d-4747-8f7c-f439b3227b11-must-gather-output\") pod \"must-gather-fmnth\" (UID: \"f8e9d6cc-a02d-4747-8f7c-f439b3227b11\") " pod="openshift-must-gather-2gd5d/must-gather-fmnth" Mar 21 05:56:32 crc kubenswrapper[4775]: I0321 05:56:32.168036 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f8e9d6cc-a02d-4747-8f7c-f439b3227b11-must-gather-output\") pod \"must-gather-fmnth\" (UID: \"f8e9d6cc-a02d-4747-8f7c-f439b3227b11\") " pod="openshift-must-gather-2gd5d/must-gather-fmnth" Mar 21 05:56:32 crc kubenswrapper[4775]: I0321 05:56:32.189548 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5jqm\" (UniqueName: \"kubernetes.io/projected/f8e9d6cc-a02d-4747-8f7c-f439b3227b11-kube-api-access-k5jqm\") pod \"must-gather-fmnth\" (UID: \"f8e9d6cc-a02d-4747-8f7c-f439b3227b11\") " pod="openshift-must-gather-2gd5d/must-gather-fmnth" Mar 21 05:56:32 crc kubenswrapper[4775]: I0321 05:56:32.362008 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2gd5d/must-gather-fmnth" Mar 21 05:56:32 crc kubenswrapper[4775]: I0321 05:56:32.481797 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:56:32 crc kubenswrapper[4775]: I0321 05:56:32.481868 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:56:33 crc kubenswrapper[4775]: I0321 05:56:33.118563 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2gd5d/must-gather-fmnth"] Mar 21 05:56:33 crc kubenswrapper[4775]: I0321 05:56:33.582380 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hvhrr" Mar 21 05:56:33 crc kubenswrapper[4775]: I0321 05:56:33.582647 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hvhrr" Mar 21 05:56:33 crc kubenswrapper[4775]: I0321 05:56:33.651323 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hvhrr" Mar 21 05:56:33 crc kubenswrapper[4775]: I0321 05:56:33.996180 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2gd5d/must-gather-fmnth" event={"ID":"f8e9d6cc-a02d-4747-8f7c-f439b3227b11","Type":"ContainerStarted","Data":"7c0d546b19a6c80ff65ea248d24b3db62171d0f126c07845976d4d5ff49b7d79"} Mar 21 05:56:33 crc kubenswrapper[4775]: I0321 05:56:33.996512 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2gd5d/must-gather-fmnth" event={"ID":"f8e9d6cc-a02d-4747-8f7c-f439b3227b11","Type":"ContainerStarted","Data":"9e98fbe03eab0e6c4baf6973b4872261d6f0c83d76001f189bcb6eab76312381"} Mar 21 05:56:33 crc kubenswrapper[4775]: I0321 05:56:33.996525 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2gd5d/must-gather-fmnth" event={"ID":"f8e9d6cc-a02d-4747-8f7c-f439b3227b11","Type":"ContainerStarted","Data":"7c4f80f0df39580d8656dd019ad5ecda89f94b6664de1b8c9c62c10087b36fe2"} Mar 21 05:56:34 crc kubenswrapper[4775]: I0321 05:56:34.021067 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-2gd5d/must-gather-fmnth" podStartSLOduration=2.021049874 podStartE2EDuration="2.021049874s" podCreationTimestamp="2026-03-21 05:56:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:56:34.013669974 +0000 UTC m=+4146.990133608" watchObservedRunningTime="2026-03-21 05:56:34.021049874 +0000 UTC m=+4146.997513498" Mar 21 05:56:34 crc kubenswrapper[4775]: I0321 05:56:34.657461 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hvhrr" Mar 21 05:56:34 crc kubenswrapper[4775]: I0321 05:56:34.713242 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hvhrr"] Mar 21 05:56:36 crc kubenswrapper[4775]: I0321 05:56:36.013017 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hvhrr" podUID="a9132264-94c1-407d-87ab-d7f546907220" containerName="registry-server" containerID="cri-o://124837a8aeed307cbf3f0d2a7f042cf5856dac1eb3286c8f4ca6d73a2fd39f43" gracePeriod=2 Mar 21 05:56:36 crc kubenswrapper[4775]: I0321 05:56:36.982091 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hvhrr" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.025830 4775 generic.go:334] "Generic (PLEG): container finished" podID="a9132264-94c1-407d-87ab-d7f546907220" containerID="124837a8aeed307cbf3f0d2a7f042cf5856dac1eb3286c8f4ca6d73a2fd39f43" exitCode=0 Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.025888 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hvhrr" event={"ID":"a9132264-94c1-407d-87ab-d7f546907220","Type":"ContainerDied","Data":"124837a8aeed307cbf3f0d2a7f042cf5856dac1eb3286c8f4ca6d73a2fd39f43"} Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.025921 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hvhrr" event={"ID":"a9132264-94c1-407d-87ab-d7f546907220","Type":"ContainerDied","Data":"2585c889d397a5cc818abfd7d487d4c812588c6d0795639c8158b4cc4e2799db"} Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.025942 4775 scope.go:117] "RemoveContainer" containerID="124837a8aeed307cbf3f0d2a7f042cf5856dac1eb3286c8f4ca6d73a2fd39f43" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.026167 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hvhrr" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.049378 4775 scope.go:117] "RemoveContainer" containerID="db2adef6aa111c2105ea410f3c602299796961892b8bbcc0eaa8ea40de324259" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.063941 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9132264-94c1-407d-87ab-d7f546907220-catalog-content\") pod \"a9132264-94c1-407d-87ab-d7f546907220\" (UID: \"a9132264-94c1-407d-87ab-d7f546907220\") " Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.064083 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9132264-94c1-407d-87ab-d7f546907220-utilities\") pod \"a9132264-94c1-407d-87ab-d7f546907220\" (UID: \"a9132264-94c1-407d-87ab-d7f546907220\") " Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.064327 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v54rl\" (UniqueName: \"kubernetes.io/projected/a9132264-94c1-407d-87ab-d7f546907220-kube-api-access-v54rl\") pod \"a9132264-94c1-407d-87ab-d7f546907220\" (UID: \"a9132264-94c1-407d-87ab-d7f546907220\") " Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.066686 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9132264-94c1-407d-87ab-d7f546907220-utilities" (OuterVolumeSpecName: "utilities") pod "a9132264-94c1-407d-87ab-d7f546907220" (UID: "a9132264-94c1-407d-87ab-d7f546907220"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.076898 4775 scope.go:117] "RemoveContainer" containerID="56ec5eaa6447799381690c7e64db1dc2298dedf7d9f2b726e9f414fe1bc074cf" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.081724 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9132264-94c1-407d-87ab-d7f546907220-kube-api-access-v54rl" (OuterVolumeSpecName: "kube-api-access-v54rl") pod "a9132264-94c1-407d-87ab-d7f546907220" (UID: "a9132264-94c1-407d-87ab-d7f546907220"). InnerVolumeSpecName "kube-api-access-v54rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.126378 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9132264-94c1-407d-87ab-d7f546907220-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a9132264-94c1-407d-87ab-d7f546907220" (UID: "a9132264-94c1-407d-87ab-d7f546907220"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.151045 4775 scope.go:117] "RemoveContainer" containerID="124837a8aeed307cbf3f0d2a7f042cf5856dac1eb3286c8f4ca6d73a2fd39f43" Mar 21 05:56:37 crc kubenswrapper[4775]: E0321 05:56:37.151630 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"124837a8aeed307cbf3f0d2a7f042cf5856dac1eb3286c8f4ca6d73a2fd39f43\": container with ID starting with 124837a8aeed307cbf3f0d2a7f042cf5856dac1eb3286c8f4ca6d73a2fd39f43 not found: ID does not exist" containerID="124837a8aeed307cbf3f0d2a7f042cf5856dac1eb3286c8f4ca6d73a2fd39f43" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.151668 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"124837a8aeed307cbf3f0d2a7f042cf5856dac1eb3286c8f4ca6d73a2fd39f43"} err="failed to get container status \"124837a8aeed307cbf3f0d2a7f042cf5856dac1eb3286c8f4ca6d73a2fd39f43\": rpc error: code = NotFound desc = could not find container \"124837a8aeed307cbf3f0d2a7f042cf5856dac1eb3286c8f4ca6d73a2fd39f43\": container with ID starting with 124837a8aeed307cbf3f0d2a7f042cf5856dac1eb3286c8f4ca6d73a2fd39f43 not found: ID does not exist" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.151695 4775 scope.go:117] "RemoveContainer" containerID="db2adef6aa111c2105ea410f3c602299796961892b8bbcc0eaa8ea40de324259" Mar 21 05:56:37 crc kubenswrapper[4775]: E0321 05:56:37.152296 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db2adef6aa111c2105ea410f3c602299796961892b8bbcc0eaa8ea40de324259\": container with ID starting with db2adef6aa111c2105ea410f3c602299796961892b8bbcc0eaa8ea40de324259 not found: ID does not exist" containerID="db2adef6aa111c2105ea410f3c602299796961892b8bbcc0eaa8ea40de324259" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.152352 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db2adef6aa111c2105ea410f3c602299796961892b8bbcc0eaa8ea40de324259"} err="failed to get container status \"db2adef6aa111c2105ea410f3c602299796961892b8bbcc0eaa8ea40de324259\": rpc error: code = NotFound desc = could not find container \"db2adef6aa111c2105ea410f3c602299796961892b8bbcc0eaa8ea40de324259\": container with ID starting with db2adef6aa111c2105ea410f3c602299796961892b8bbcc0eaa8ea40de324259 not found: ID does not exist" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.152385 4775 scope.go:117] "RemoveContainer" containerID="56ec5eaa6447799381690c7e64db1dc2298dedf7d9f2b726e9f414fe1bc074cf" Mar 21 05:56:37 crc kubenswrapper[4775]: E0321 05:56:37.152731 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56ec5eaa6447799381690c7e64db1dc2298dedf7d9f2b726e9f414fe1bc074cf\": container with ID starting with 56ec5eaa6447799381690c7e64db1dc2298dedf7d9f2b726e9f414fe1bc074cf not found: ID does not exist" containerID="56ec5eaa6447799381690c7e64db1dc2298dedf7d9f2b726e9f414fe1bc074cf" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.152766 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56ec5eaa6447799381690c7e64db1dc2298dedf7d9f2b726e9f414fe1bc074cf"} err="failed to get container status \"56ec5eaa6447799381690c7e64db1dc2298dedf7d9f2b726e9f414fe1bc074cf\": rpc error: code = NotFound desc = could not find container \"56ec5eaa6447799381690c7e64db1dc2298dedf7d9f2b726e9f414fe1bc074cf\": container with ID starting with 56ec5eaa6447799381690c7e64db1dc2298dedf7d9f2b726e9f414fe1bc074cf not found: ID does not exist" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.167356 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9132264-94c1-407d-87ab-d7f546907220-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.167409 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v54rl\" (UniqueName: \"kubernetes.io/projected/a9132264-94c1-407d-87ab-d7f546907220-kube-api-access-v54rl\") on node \"crc\" DevicePath \"\"" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.167422 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9132264-94c1-407d-87ab-d7f546907220-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.319947 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2gd5d/crc-debug-d75fs"] Mar 21 05:56:37 crc kubenswrapper[4775]: E0321 05:56:37.320394 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9132264-94c1-407d-87ab-d7f546907220" containerName="extract-utilities" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.320413 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9132264-94c1-407d-87ab-d7f546907220" containerName="extract-utilities" Mar 21 05:56:37 crc kubenswrapper[4775]: E0321 05:56:37.320430 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9132264-94c1-407d-87ab-d7f546907220" containerName="extract-content" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.320439 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9132264-94c1-407d-87ab-d7f546907220" containerName="extract-content" Mar 21 05:56:37 crc kubenswrapper[4775]: E0321 05:56:37.320458 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9132264-94c1-407d-87ab-d7f546907220" containerName="registry-server" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.320465 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9132264-94c1-407d-87ab-d7f546907220" containerName="registry-server" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.320829 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9132264-94c1-407d-87ab-d7f546907220" containerName="registry-server" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.321494 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2gd5d/crc-debug-d75fs" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.365502 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hvhrr"] Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.376087 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hvhrr"] Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.473375 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl9hg\" (UniqueName: \"kubernetes.io/projected/02fd708d-0508-40e9-8591-02cc73141ca4-kube-api-access-hl9hg\") pod \"crc-debug-d75fs\" (UID: \"02fd708d-0508-40e9-8591-02cc73141ca4\") " pod="openshift-must-gather-2gd5d/crc-debug-d75fs" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.473428 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/02fd708d-0508-40e9-8591-02cc73141ca4-host\") pod \"crc-debug-d75fs\" (UID: \"02fd708d-0508-40e9-8591-02cc73141ca4\") " pod="openshift-must-gather-2gd5d/crc-debug-d75fs" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.575691 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl9hg\" (UniqueName: \"kubernetes.io/projected/02fd708d-0508-40e9-8591-02cc73141ca4-kube-api-access-hl9hg\") pod \"crc-debug-d75fs\" (UID: \"02fd708d-0508-40e9-8591-02cc73141ca4\") " pod="openshift-must-gather-2gd5d/crc-debug-d75fs" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.575759 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/02fd708d-0508-40e9-8591-02cc73141ca4-host\") pod \"crc-debug-d75fs\" (UID: \"02fd708d-0508-40e9-8591-02cc73141ca4\") " pod="openshift-must-gather-2gd5d/crc-debug-d75fs" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.575984 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/02fd708d-0508-40e9-8591-02cc73141ca4-host\") pod \"crc-debug-d75fs\" (UID: \"02fd708d-0508-40e9-8591-02cc73141ca4\") " pod="openshift-must-gather-2gd5d/crc-debug-d75fs" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.598175 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl9hg\" (UniqueName: \"kubernetes.io/projected/02fd708d-0508-40e9-8591-02cc73141ca4-kube-api-access-hl9hg\") pod \"crc-debug-d75fs\" (UID: \"02fd708d-0508-40e9-8591-02cc73141ca4\") " pod="openshift-must-gather-2gd5d/crc-debug-d75fs" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.643399 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2gd5d/crc-debug-d75fs" Mar 21 05:56:37 crc kubenswrapper[4775]: I0321 05:56:37.675505 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9132264-94c1-407d-87ab-d7f546907220" path="/var/lib/kubelet/pods/a9132264-94c1-407d-87ab-d7f546907220/volumes" Mar 21 05:56:38 crc kubenswrapper[4775]: I0321 05:56:38.050409 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2gd5d/crc-debug-d75fs" event={"ID":"02fd708d-0508-40e9-8591-02cc73141ca4","Type":"ContainerStarted","Data":"bd3bbff9eadc760f7629d28db8d7131fe45713ac32e2244fcd75d129dcb618b6"} Mar 21 05:56:39 crc kubenswrapper[4775]: I0321 05:56:39.062908 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2gd5d/crc-debug-d75fs" event={"ID":"02fd708d-0508-40e9-8591-02cc73141ca4","Type":"ContainerStarted","Data":"6458d3534ba667913e8d34c99b54452e526574493e208030a9412e60968bc790"} Mar 21 05:56:39 crc kubenswrapper[4775]: I0321 05:56:39.081345 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-2gd5d/crc-debug-d75fs" podStartSLOduration=2.0813220550000002 podStartE2EDuration="2.081322055s" podCreationTimestamp="2026-03-21 05:56:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 05:56:39.079770141 +0000 UTC m=+4152.056233775" watchObservedRunningTime="2026-03-21 05:56:39.081322055 +0000 UTC m=+4152.057785679" Mar 21 05:56:49 crc kubenswrapper[4775]: I0321 05:56:49.170611 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hf79m"] Mar 21 05:56:49 crc kubenswrapper[4775]: I0321 05:56:49.173991 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hf79m" Mar 21 05:56:49 crc kubenswrapper[4775]: I0321 05:56:49.179823 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hf79m"] Mar 21 05:56:49 crc kubenswrapper[4775]: I0321 05:56:49.309280 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc7n2\" (UniqueName: \"kubernetes.io/projected/8022fa19-b95f-4ac2-a8cc-6a47a874fd34-kube-api-access-bc7n2\") pod \"redhat-marketplace-hf79m\" (UID: \"8022fa19-b95f-4ac2-a8cc-6a47a874fd34\") " pod="openshift-marketplace/redhat-marketplace-hf79m" Mar 21 05:56:49 crc kubenswrapper[4775]: I0321 05:56:49.309339 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8022fa19-b95f-4ac2-a8cc-6a47a874fd34-utilities\") pod \"redhat-marketplace-hf79m\" (UID: \"8022fa19-b95f-4ac2-a8cc-6a47a874fd34\") " pod="openshift-marketplace/redhat-marketplace-hf79m" Mar 21 05:56:49 crc kubenswrapper[4775]: I0321 05:56:49.309765 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8022fa19-b95f-4ac2-a8cc-6a47a874fd34-catalog-content\") pod \"redhat-marketplace-hf79m\" (UID: \"8022fa19-b95f-4ac2-a8cc-6a47a874fd34\") " pod="openshift-marketplace/redhat-marketplace-hf79m" Mar 21 05:56:49 crc kubenswrapper[4775]: I0321 05:56:49.411791 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc7n2\" (UniqueName: \"kubernetes.io/projected/8022fa19-b95f-4ac2-a8cc-6a47a874fd34-kube-api-access-bc7n2\") pod \"redhat-marketplace-hf79m\" (UID: \"8022fa19-b95f-4ac2-a8cc-6a47a874fd34\") " pod="openshift-marketplace/redhat-marketplace-hf79m" Mar 21 05:56:49 crc kubenswrapper[4775]: I0321 05:56:49.411864 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8022fa19-b95f-4ac2-a8cc-6a47a874fd34-utilities\") pod \"redhat-marketplace-hf79m\" (UID: \"8022fa19-b95f-4ac2-a8cc-6a47a874fd34\") " pod="openshift-marketplace/redhat-marketplace-hf79m" Mar 21 05:56:49 crc kubenswrapper[4775]: I0321 05:56:49.412019 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8022fa19-b95f-4ac2-a8cc-6a47a874fd34-catalog-content\") pod \"redhat-marketplace-hf79m\" (UID: \"8022fa19-b95f-4ac2-a8cc-6a47a874fd34\") " pod="openshift-marketplace/redhat-marketplace-hf79m" Mar 21 05:56:49 crc kubenswrapper[4775]: I0321 05:56:49.412436 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8022fa19-b95f-4ac2-a8cc-6a47a874fd34-utilities\") pod \"redhat-marketplace-hf79m\" (UID: \"8022fa19-b95f-4ac2-a8cc-6a47a874fd34\") " pod="openshift-marketplace/redhat-marketplace-hf79m" Mar 21 05:56:49 crc kubenswrapper[4775]: I0321 05:56:49.412651 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8022fa19-b95f-4ac2-a8cc-6a47a874fd34-catalog-content\") pod \"redhat-marketplace-hf79m\" (UID: \"8022fa19-b95f-4ac2-a8cc-6a47a874fd34\") " pod="openshift-marketplace/redhat-marketplace-hf79m" Mar 21 05:56:49 crc kubenswrapper[4775]: I0321 05:56:49.439071 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc7n2\" (UniqueName: \"kubernetes.io/projected/8022fa19-b95f-4ac2-a8cc-6a47a874fd34-kube-api-access-bc7n2\") pod \"redhat-marketplace-hf79m\" (UID: \"8022fa19-b95f-4ac2-a8cc-6a47a874fd34\") " pod="openshift-marketplace/redhat-marketplace-hf79m" Mar 21 05:56:49 crc kubenswrapper[4775]: I0321 05:56:49.502984 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hf79m" Mar 21 05:56:49 crc kubenswrapper[4775]: I0321 05:56:49.580816 4775 scope.go:117] "RemoveContainer" containerID="21e7854b61e199821209ba3f9487dda8cf82083673134efff058e8fd96d17f3f" Mar 21 05:56:50 crc kubenswrapper[4775]: I0321 05:56:50.098556 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hf79m"] Mar 21 05:56:50 crc kubenswrapper[4775]: I0321 05:56:50.169342 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hf79m" event={"ID":"8022fa19-b95f-4ac2-a8cc-6a47a874fd34","Type":"ContainerStarted","Data":"6119ecf3284a2c38c85c0126ae363c2d0c4d602c9e56e0ece76b212ea792f2ac"} Mar 21 05:56:51 crc kubenswrapper[4775]: I0321 05:56:51.180987 4775 generic.go:334] "Generic (PLEG): container finished" podID="8022fa19-b95f-4ac2-a8cc-6a47a874fd34" containerID="b49b1a271ad4bd5b721f831a551f4d151d0600192cfd36b385e5c58ea9cc4d89" exitCode=0 Mar 21 05:56:51 crc kubenswrapper[4775]: I0321 05:56:51.181096 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hf79m" event={"ID":"8022fa19-b95f-4ac2-a8cc-6a47a874fd34","Type":"ContainerDied","Data":"b49b1a271ad4bd5b721f831a551f4d151d0600192cfd36b385e5c58ea9cc4d89"} Mar 21 05:56:52 crc kubenswrapper[4775]: I0321 05:56:52.192572 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hf79m" event={"ID":"8022fa19-b95f-4ac2-a8cc-6a47a874fd34","Type":"ContainerStarted","Data":"26d7e2528fec25cd60be964aaf6b1f8773399257833c635b71323a506857fd37"} Mar 21 05:56:54 crc kubenswrapper[4775]: I0321 05:56:54.225545 4775 generic.go:334] "Generic (PLEG): container finished" podID="8022fa19-b95f-4ac2-a8cc-6a47a874fd34" containerID="26d7e2528fec25cd60be964aaf6b1f8773399257833c635b71323a506857fd37" exitCode=0 Mar 21 05:56:54 crc kubenswrapper[4775]: I0321 05:56:54.225632 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hf79m" event={"ID":"8022fa19-b95f-4ac2-a8cc-6a47a874fd34","Type":"ContainerDied","Data":"26d7e2528fec25cd60be964aaf6b1f8773399257833c635b71323a506857fd37"} Mar 21 05:56:56 crc kubenswrapper[4775]: I0321 05:56:56.252763 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hf79m" event={"ID":"8022fa19-b95f-4ac2-a8cc-6a47a874fd34","Type":"ContainerStarted","Data":"388eee5beefb800def090d2007d8dfc0e1863d473480f91a687b60b9e0f1893c"} Mar 21 05:56:56 crc kubenswrapper[4775]: I0321 05:56:56.280717 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hf79m" podStartSLOduration=3.448545967 podStartE2EDuration="7.280693148s" podCreationTimestamp="2026-03-21 05:56:49 +0000 UTC" firstStartedPulling="2026-03-21 05:56:51.184549508 +0000 UTC m=+4164.161013132" lastFinishedPulling="2026-03-21 05:56:55.016696689 +0000 UTC m=+4167.993160313" observedRunningTime="2026-03-21 05:56:56.275880162 +0000 UTC m=+4169.252343796" watchObservedRunningTime="2026-03-21 05:56:56.280693148 +0000 UTC m=+4169.257156782" Mar 21 05:56:59 crc kubenswrapper[4775]: I0321 05:56:59.503944 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hf79m" Mar 21 05:56:59 crc kubenswrapper[4775]: I0321 05:56:59.504413 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hf79m" Mar 21 05:56:59 crc kubenswrapper[4775]: I0321 05:56:59.557351 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hf79m" Mar 21 05:57:00 crc kubenswrapper[4775]: I0321 05:57:00.337895 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hf79m" Mar 21 05:57:00 crc kubenswrapper[4775]: I0321 05:57:00.408220 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hf79m"] Mar 21 05:57:02 crc kubenswrapper[4775]: I0321 05:57:02.301218 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hf79m" podUID="8022fa19-b95f-4ac2-a8cc-6a47a874fd34" containerName="registry-server" containerID="cri-o://388eee5beefb800def090d2007d8dfc0e1863d473480f91a687b60b9e0f1893c" gracePeriod=2 Mar 21 05:57:02 crc kubenswrapper[4775]: I0321 05:57:02.482595 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:57:02 crc kubenswrapper[4775]: I0321 05:57:02.482650 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:57:02 crc kubenswrapper[4775]: I0321 05:57:02.875416 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hf79m" Mar 21 05:57:02 crc kubenswrapper[4775]: I0321 05:57:02.927045 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8022fa19-b95f-4ac2-a8cc-6a47a874fd34-utilities" (OuterVolumeSpecName: "utilities") pod "8022fa19-b95f-4ac2-a8cc-6a47a874fd34" (UID: "8022fa19-b95f-4ac2-a8cc-6a47a874fd34"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:57:02 crc kubenswrapper[4775]: I0321 05:57:02.927418 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8022fa19-b95f-4ac2-a8cc-6a47a874fd34-utilities\") pod \"8022fa19-b95f-4ac2-a8cc-6a47a874fd34\" (UID: \"8022fa19-b95f-4ac2-a8cc-6a47a874fd34\") " Mar 21 05:57:02 crc kubenswrapper[4775]: I0321 05:57:02.927597 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8022fa19-b95f-4ac2-a8cc-6a47a874fd34-catalog-content\") pod \"8022fa19-b95f-4ac2-a8cc-6a47a874fd34\" (UID: \"8022fa19-b95f-4ac2-a8cc-6a47a874fd34\") " Mar 21 05:57:02 crc kubenswrapper[4775]: I0321 05:57:02.927651 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bc7n2\" (UniqueName: \"kubernetes.io/projected/8022fa19-b95f-4ac2-a8cc-6a47a874fd34-kube-api-access-bc7n2\") pod \"8022fa19-b95f-4ac2-a8cc-6a47a874fd34\" (UID: \"8022fa19-b95f-4ac2-a8cc-6a47a874fd34\") " Mar 21 05:57:02 crc kubenswrapper[4775]: I0321 05:57:02.928056 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8022fa19-b95f-4ac2-a8cc-6a47a874fd34-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 05:57:02 crc kubenswrapper[4775]: I0321 05:57:02.963008 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8022fa19-b95f-4ac2-a8cc-6a47a874fd34-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8022fa19-b95f-4ac2-a8cc-6a47a874fd34" (UID: "8022fa19-b95f-4ac2-a8cc-6a47a874fd34"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 05:57:03 crc kubenswrapper[4775]: I0321 05:57:03.029695 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8022fa19-b95f-4ac2-a8cc-6a47a874fd34-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 05:57:03 crc kubenswrapper[4775]: I0321 05:57:03.312722 4775 generic.go:334] "Generic (PLEG): container finished" podID="8022fa19-b95f-4ac2-a8cc-6a47a874fd34" containerID="388eee5beefb800def090d2007d8dfc0e1863d473480f91a687b60b9e0f1893c" exitCode=0 Mar 21 05:57:03 crc kubenswrapper[4775]: I0321 05:57:03.312766 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hf79m" event={"ID":"8022fa19-b95f-4ac2-a8cc-6a47a874fd34","Type":"ContainerDied","Data":"388eee5beefb800def090d2007d8dfc0e1863d473480f91a687b60b9e0f1893c"} Mar 21 05:57:03 crc kubenswrapper[4775]: I0321 05:57:03.312799 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hf79m" Mar 21 05:57:03 crc kubenswrapper[4775]: I0321 05:57:03.312824 4775 scope.go:117] "RemoveContainer" containerID="388eee5beefb800def090d2007d8dfc0e1863d473480f91a687b60b9e0f1893c" Mar 21 05:57:03 crc kubenswrapper[4775]: I0321 05:57:03.312810 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hf79m" event={"ID":"8022fa19-b95f-4ac2-a8cc-6a47a874fd34","Type":"ContainerDied","Data":"6119ecf3284a2c38c85c0126ae363c2d0c4d602c9e56e0ece76b212ea792f2ac"} Mar 21 05:57:03 crc kubenswrapper[4775]: I0321 05:57:03.338412 4775 scope.go:117] "RemoveContainer" containerID="26d7e2528fec25cd60be964aaf6b1f8773399257833c635b71323a506857fd37" Mar 21 05:57:03 crc kubenswrapper[4775]: I0321 05:57:03.500384 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8022fa19-b95f-4ac2-a8cc-6a47a874fd34-kube-api-access-bc7n2" (OuterVolumeSpecName: "kube-api-access-bc7n2") pod "8022fa19-b95f-4ac2-a8cc-6a47a874fd34" (UID: "8022fa19-b95f-4ac2-a8cc-6a47a874fd34"). InnerVolumeSpecName "kube-api-access-bc7n2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:57:03 crc kubenswrapper[4775]: I0321 05:57:03.537742 4775 scope.go:117] "RemoveContainer" containerID="b49b1a271ad4bd5b721f831a551f4d151d0600192cfd36b385e5c58ea9cc4d89" Mar 21 05:57:03 crc kubenswrapper[4775]: I0321 05:57:03.538507 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bc7n2\" (UniqueName: \"kubernetes.io/projected/8022fa19-b95f-4ac2-a8cc-6a47a874fd34-kube-api-access-bc7n2\") on node \"crc\" DevicePath \"\"" Mar 21 05:57:03 crc kubenswrapper[4775]: I0321 05:57:03.614626 4775 scope.go:117] "RemoveContainer" containerID="388eee5beefb800def090d2007d8dfc0e1863d473480f91a687b60b9e0f1893c" Mar 21 05:57:03 crc kubenswrapper[4775]: E0321 05:57:03.615052 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"388eee5beefb800def090d2007d8dfc0e1863d473480f91a687b60b9e0f1893c\": container with ID starting with 388eee5beefb800def090d2007d8dfc0e1863d473480f91a687b60b9e0f1893c not found: ID does not exist" containerID="388eee5beefb800def090d2007d8dfc0e1863d473480f91a687b60b9e0f1893c" Mar 21 05:57:03 crc kubenswrapper[4775]: I0321 05:57:03.615100 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"388eee5beefb800def090d2007d8dfc0e1863d473480f91a687b60b9e0f1893c"} err="failed to get container status \"388eee5beefb800def090d2007d8dfc0e1863d473480f91a687b60b9e0f1893c\": rpc error: code = NotFound desc = could not find container \"388eee5beefb800def090d2007d8dfc0e1863d473480f91a687b60b9e0f1893c\": container with ID starting with 388eee5beefb800def090d2007d8dfc0e1863d473480f91a687b60b9e0f1893c not found: ID does not exist" Mar 21 05:57:03 crc kubenswrapper[4775]: I0321 05:57:03.615142 4775 scope.go:117] "RemoveContainer" containerID="26d7e2528fec25cd60be964aaf6b1f8773399257833c635b71323a506857fd37" Mar 21 05:57:03 crc kubenswrapper[4775]: E0321 05:57:03.615451 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26d7e2528fec25cd60be964aaf6b1f8773399257833c635b71323a506857fd37\": container with ID starting with 26d7e2528fec25cd60be964aaf6b1f8773399257833c635b71323a506857fd37 not found: ID does not exist" containerID="26d7e2528fec25cd60be964aaf6b1f8773399257833c635b71323a506857fd37" Mar 21 05:57:03 crc kubenswrapper[4775]: I0321 05:57:03.615582 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26d7e2528fec25cd60be964aaf6b1f8773399257833c635b71323a506857fd37"} err="failed to get container status \"26d7e2528fec25cd60be964aaf6b1f8773399257833c635b71323a506857fd37\": rpc error: code = NotFound desc = could not find container \"26d7e2528fec25cd60be964aaf6b1f8773399257833c635b71323a506857fd37\": container with ID starting with 26d7e2528fec25cd60be964aaf6b1f8773399257833c635b71323a506857fd37 not found: ID does not exist" Mar 21 05:57:03 crc kubenswrapper[4775]: I0321 05:57:03.615916 4775 scope.go:117] "RemoveContainer" containerID="b49b1a271ad4bd5b721f831a551f4d151d0600192cfd36b385e5c58ea9cc4d89" Mar 21 05:57:03 crc kubenswrapper[4775]: E0321 05:57:03.616398 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b49b1a271ad4bd5b721f831a551f4d151d0600192cfd36b385e5c58ea9cc4d89\": container with ID starting with b49b1a271ad4bd5b721f831a551f4d151d0600192cfd36b385e5c58ea9cc4d89 not found: ID does not exist" containerID="b49b1a271ad4bd5b721f831a551f4d151d0600192cfd36b385e5c58ea9cc4d89" Mar 21 05:57:03 crc kubenswrapper[4775]: I0321 05:57:03.616435 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b49b1a271ad4bd5b721f831a551f4d151d0600192cfd36b385e5c58ea9cc4d89"} err="failed to get container status \"b49b1a271ad4bd5b721f831a551f4d151d0600192cfd36b385e5c58ea9cc4d89\": rpc error: code = NotFound desc = could not find container \"b49b1a271ad4bd5b721f831a551f4d151d0600192cfd36b385e5c58ea9cc4d89\": container with ID starting with b49b1a271ad4bd5b721f831a551f4d151d0600192cfd36b385e5c58ea9cc4d89 not found: ID does not exist" Mar 21 05:57:03 crc kubenswrapper[4775]: I0321 05:57:03.682052 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hf79m"] Mar 21 05:57:03 crc kubenswrapper[4775]: I0321 05:57:03.707515 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hf79m"] Mar 21 05:57:05 crc kubenswrapper[4775]: I0321 05:57:05.674032 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8022fa19-b95f-4ac2-a8cc-6a47a874fd34" path="/var/lib/kubelet/pods/8022fa19-b95f-4ac2-a8cc-6a47a874fd34/volumes" Mar 21 05:57:16 crc kubenswrapper[4775]: I0321 05:57:16.448886 4775 generic.go:334] "Generic (PLEG): container finished" podID="02fd708d-0508-40e9-8591-02cc73141ca4" containerID="6458d3534ba667913e8d34c99b54452e526574493e208030a9412e60968bc790" exitCode=0 Mar 21 05:57:16 crc kubenswrapper[4775]: I0321 05:57:16.448989 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2gd5d/crc-debug-d75fs" event={"ID":"02fd708d-0508-40e9-8591-02cc73141ca4","Type":"ContainerDied","Data":"6458d3534ba667913e8d34c99b54452e526574493e208030a9412e60968bc790"} Mar 21 05:57:17 crc kubenswrapper[4775]: I0321 05:57:17.589211 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2gd5d/crc-debug-d75fs" Mar 21 05:57:17 crc kubenswrapper[4775]: I0321 05:57:17.602431 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hl9hg\" (UniqueName: \"kubernetes.io/projected/02fd708d-0508-40e9-8591-02cc73141ca4-kube-api-access-hl9hg\") pod \"02fd708d-0508-40e9-8591-02cc73141ca4\" (UID: \"02fd708d-0508-40e9-8591-02cc73141ca4\") " Mar 21 05:57:17 crc kubenswrapper[4775]: I0321 05:57:17.602547 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/02fd708d-0508-40e9-8591-02cc73141ca4-host\") pod \"02fd708d-0508-40e9-8591-02cc73141ca4\" (UID: \"02fd708d-0508-40e9-8591-02cc73141ca4\") " Mar 21 05:57:17 crc kubenswrapper[4775]: I0321 05:57:17.602952 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02fd708d-0508-40e9-8591-02cc73141ca4-host" (OuterVolumeSpecName: "host") pod "02fd708d-0508-40e9-8591-02cc73141ca4" (UID: "02fd708d-0508-40e9-8591-02cc73141ca4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:57:17 crc kubenswrapper[4775]: I0321 05:57:17.628319 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02fd708d-0508-40e9-8591-02cc73141ca4-kube-api-access-hl9hg" (OuterVolumeSpecName: "kube-api-access-hl9hg") pod "02fd708d-0508-40e9-8591-02cc73141ca4" (UID: "02fd708d-0508-40e9-8591-02cc73141ca4"). InnerVolumeSpecName "kube-api-access-hl9hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:57:17 crc kubenswrapper[4775]: I0321 05:57:17.636423 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2gd5d/crc-debug-d75fs"] Mar 21 05:57:17 crc kubenswrapper[4775]: I0321 05:57:17.648730 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2gd5d/crc-debug-d75fs"] Mar 21 05:57:17 crc kubenswrapper[4775]: I0321 05:57:17.674376 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02fd708d-0508-40e9-8591-02cc73141ca4" path="/var/lib/kubelet/pods/02fd708d-0508-40e9-8591-02cc73141ca4/volumes" Mar 21 05:57:17 crc kubenswrapper[4775]: I0321 05:57:17.714830 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hl9hg\" (UniqueName: \"kubernetes.io/projected/02fd708d-0508-40e9-8591-02cc73141ca4-kube-api-access-hl9hg\") on node \"crc\" DevicePath \"\"" Mar 21 05:57:17 crc kubenswrapper[4775]: I0321 05:57:17.714870 4775 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/02fd708d-0508-40e9-8591-02cc73141ca4-host\") on node \"crc\" DevicePath \"\"" Mar 21 05:57:18 crc kubenswrapper[4775]: I0321 05:57:18.476590 4775 scope.go:117] "RemoveContainer" containerID="6458d3534ba667913e8d34c99b54452e526574493e208030a9412e60968bc790" Mar 21 05:57:18 crc kubenswrapper[4775]: I0321 05:57:18.476818 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2gd5d/crc-debug-d75fs" Mar 21 05:57:18 crc kubenswrapper[4775]: I0321 05:57:18.932980 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2gd5d/crc-debug-8rplg"] Mar 21 05:57:18 crc kubenswrapper[4775]: E0321 05:57:18.933787 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8022fa19-b95f-4ac2-a8cc-6a47a874fd34" containerName="extract-content" Mar 21 05:57:18 crc kubenswrapper[4775]: I0321 05:57:18.933799 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8022fa19-b95f-4ac2-a8cc-6a47a874fd34" containerName="extract-content" Mar 21 05:57:18 crc kubenswrapper[4775]: E0321 05:57:18.933815 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02fd708d-0508-40e9-8591-02cc73141ca4" containerName="container-00" Mar 21 05:57:18 crc kubenswrapper[4775]: I0321 05:57:18.933823 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="02fd708d-0508-40e9-8591-02cc73141ca4" containerName="container-00" Mar 21 05:57:18 crc kubenswrapper[4775]: E0321 05:57:18.933843 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8022fa19-b95f-4ac2-a8cc-6a47a874fd34" containerName="registry-server" Mar 21 05:57:18 crc kubenswrapper[4775]: I0321 05:57:18.933851 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8022fa19-b95f-4ac2-a8cc-6a47a874fd34" containerName="registry-server" Mar 21 05:57:18 crc kubenswrapper[4775]: E0321 05:57:18.933871 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8022fa19-b95f-4ac2-a8cc-6a47a874fd34" containerName="extract-utilities" Mar 21 05:57:18 crc kubenswrapper[4775]: I0321 05:57:18.933876 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8022fa19-b95f-4ac2-a8cc-6a47a874fd34" containerName="extract-utilities" Mar 21 05:57:18 crc kubenswrapper[4775]: I0321 05:57:18.934033 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8022fa19-b95f-4ac2-a8cc-6a47a874fd34" containerName="registry-server" Mar 21 05:57:18 crc kubenswrapper[4775]: I0321 05:57:18.934055 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="02fd708d-0508-40e9-8591-02cc73141ca4" containerName="container-00" Mar 21 05:57:18 crc kubenswrapper[4775]: I0321 05:57:18.935005 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2gd5d/crc-debug-8rplg" Mar 21 05:57:19 crc kubenswrapper[4775]: I0321 05:57:19.033902 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90863e29-81bb-477f-846a-7f9c27ed3b89-host\") pod \"crc-debug-8rplg\" (UID: \"90863e29-81bb-477f-846a-7f9c27ed3b89\") " pod="openshift-must-gather-2gd5d/crc-debug-8rplg" Mar 21 05:57:19 crc kubenswrapper[4775]: I0321 05:57:19.034351 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlrj8\" (UniqueName: \"kubernetes.io/projected/90863e29-81bb-477f-846a-7f9c27ed3b89-kube-api-access-mlrj8\") pod \"crc-debug-8rplg\" (UID: \"90863e29-81bb-477f-846a-7f9c27ed3b89\") " pod="openshift-must-gather-2gd5d/crc-debug-8rplg" Mar 21 05:57:19 crc kubenswrapper[4775]: I0321 05:57:19.135849 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlrj8\" (UniqueName: \"kubernetes.io/projected/90863e29-81bb-477f-846a-7f9c27ed3b89-kube-api-access-mlrj8\") pod \"crc-debug-8rplg\" (UID: \"90863e29-81bb-477f-846a-7f9c27ed3b89\") " pod="openshift-must-gather-2gd5d/crc-debug-8rplg" Mar 21 05:57:19 crc kubenswrapper[4775]: I0321 05:57:19.136311 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90863e29-81bb-477f-846a-7f9c27ed3b89-host\") pod \"crc-debug-8rplg\" (UID: \"90863e29-81bb-477f-846a-7f9c27ed3b89\") " pod="openshift-must-gather-2gd5d/crc-debug-8rplg" Mar 21 05:57:19 crc kubenswrapper[4775]: I0321 05:57:19.136425 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90863e29-81bb-477f-846a-7f9c27ed3b89-host\") pod \"crc-debug-8rplg\" (UID: \"90863e29-81bb-477f-846a-7f9c27ed3b89\") " pod="openshift-must-gather-2gd5d/crc-debug-8rplg" Mar 21 05:57:19 crc kubenswrapper[4775]: I0321 05:57:19.160251 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlrj8\" (UniqueName: \"kubernetes.io/projected/90863e29-81bb-477f-846a-7f9c27ed3b89-kube-api-access-mlrj8\") pod \"crc-debug-8rplg\" (UID: \"90863e29-81bb-477f-846a-7f9c27ed3b89\") " pod="openshift-must-gather-2gd5d/crc-debug-8rplg" Mar 21 05:57:19 crc kubenswrapper[4775]: I0321 05:57:19.291464 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2gd5d/crc-debug-8rplg" Mar 21 05:57:19 crc kubenswrapper[4775]: I0321 05:57:19.486670 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2gd5d/crc-debug-8rplg" event={"ID":"90863e29-81bb-477f-846a-7f9c27ed3b89","Type":"ContainerStarted","Data":"a273f01b54aa2eb40ad9286487d5ace47db2a7af3a064d365f343f92de6cf798"} Mar 21 05:57:20 crc kubenswrapper[4775]: I0321 05:57:20.506976 4775 generic.go:334] "Generic (PLEG): container finished" podID="90863e29-81bb-477f-846a-7f9c27ed3b89" containerID="9b324351e03c9b33d2013ebe58b0174e6c17561e8f6660c56e5da466bb4b8c3d" exitCode=0 Mar 21 05:57:20 crc kubenswrapper[4775]: I0321 05:57:20.507693 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2gd5d/crc-debug-8rplg" event={"ID":"90863e29-81bb-477f-846a-7f9c27ed3b89","Type":"ContainerDied","Data":"9b324351e03c9b33d2013ebe58b0174e6c17561e8f6660c56e5da466bb4b8c3d"} Mar 21 05:57:20 crc kubenswrapper[4775]: I0321 05:57:20.980326 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2gd5d/crc-debug-8rplg"] Mar 21 05:57:20 crc kubenswrapper[4775]: I0321 05:57:20.994145 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2gd5d/crc-debug-8rplg"] Mar 21 05:57:21 crc kubenswrapper[4775]: I0321 05:57:21.658456 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2gd5d/crc-debug-8rplg" Mar 21 05:57:21 crc kubenswrapper[4775]: I0321 05:57:21.792376 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90863e29-81bb-477f-846a-7f9c27ed3b89-host\") pod \"90863e29-81bb-477f-846a-7f9c27ed3b89\" (UID: \"90863e29-81bb-477f-846a-7f9c27ed3b89\") " Mar 21 05:57:21 crc kubenswrapper[4775]: I0321 05:57:21.792467 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/90863e29-81bb-477f-846a-7f9c27ed3b89-host" (OuterVolumeSpecName: "host") pod "90863e29-81bb-477f-846a-7f9c27ed3b89" (UID: "90863e29-81bb-477f-846a-7f9c27ed3b89"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:57:21 crc kubenswrapper[4775]: I0321 05:57:21.792873 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlrj8\" (UniqueName: \"kubernetes.io/projected/90863e29-81bb-477f-846a-7f9c27ed3b89-kube-api-access-mlrj8\") pod \"90863e29-81bb-477f-846a-7f9c27ed3b89\" (UID: \"90863e29-81bb-477f-846a-7f9c27ed3b89\") " Mar 21 05:57:21 crc kubenswrapper[4775]: I0321 05:57:21.793337 4775 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90863e29-81bb-477f-846a-7f9c27ed3b89-host\") on node \"crc\" DevicePath \"\"" Mar 21 05:57:21 crc kubenswrapper[4775]: I0321 05:57:21.802441 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90863e29-81bb-477f-846a-7f9c27ed3b89-kube-api-access-mlrj8" (OuterVolumeSpecName: "kube-api-access-mlrj8") pod "90863e29-81bb-477f-846a-7f9c27ed3b89" (UID: "90863e29-81bb-477f-846a-7f9c27ed3b89"). InnerVolumeSpecName "kube-api-access-mlrj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:57:21 crc kubenswrapper[4775]: I0321 05:57:21.895862 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlrj8\" (UniqueName: \"kubernetes.io/projected/90863e29-81bb-477f-846a-7f9c27ed3b89-kube-api-access-mlrj8\") on node \"crc\" DevicePath \"\"" Mar 21 05:57:22 crc kubenswrapper[4775]: I0321 05:57:22.188557 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2gd5d/crc-debug-xh575"] Mar 21 05:57:22 crc kubenswrapper[4775]: E0321 05:57:22.189099 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90863e29-81bb-477f-846a-7f9c27ed3b89" containerName="container-00" Mar 21 05:57:22 crc kubenswrapper[4775]: I0321 05:57:22.189138 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="90863e29-81bb-477f-846a-7f9c27ed3b89" containerName="container-00" Mar 21 05:57:22 crc kubenswrapper[4775]: I0321 05:57:22.189351 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="90863e29-81bb-477f-846a-7f9c27ed3b89" containerName="container-00" Mar 21 05:57:22 crc kubenswrapper[4775]: I0321 05:57:22.190148 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2gd5d/crc-debug-xh575" Mar 21 05:57:22 crc kubenswrapper[4775]: I0321 05:57:22.207240 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6f7e5b9b-4224-400b-b466-98e72682898e-host\") pod \"crc-debug-xh575\" (UID: \"6f7e5b9b-4224-400b-b466-98e72682898e\") " pod="openshift-must-gather-2gd5d/crc-debug-xh575" Mar 21 05:57:22 crc kubenswrapper[4775]: I0321 05:57:22.207432 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q2s5\" (UniqueName: \"kubernetes.io/projected/6f7e5b9b-4224-400b-b466-98e72682898e-kube-api-access-6q2s5\") pod \"crc-debug-xh575\" (UID: \"6f7e5b9b-4224-400b-b466-98e72682898e\") " pod="openshift-must-gather-2gd5d/crc-debug-xh575" Mar 21 05:57:22 crc kubenswrapper[4775]: I0321 05:57:22.308188 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6f7e5b9b-4224-400b-b466-98e72682898e-host\") pod \"crc-debug-xh575\" (UID: \"6f7e5b9b-4224-400b-b466-98e72682898e\") " pod="openshift-must-gather-2gd5d/crc-debug-xh575" Mar 21 05:57:22 crc kubenswrapper[4775]: I0321 05:57:22.308361 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6f7e5b9b-4224-400b-b466-98e72682898e-host\") pod \"crc-debug-xh575\" (UID: \"6f7e5b9b-4224-400b-b466-98e72682898e\") " pod="openshift-must-gather-2gd5d/crc-debug-xh575" Mar 21 05:57:22 crc kubenswrapper[4775]: I0321 05:57:22.308421 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q2s5\" (UniqueName: \"kubernetes.io/projected/6f7e5b9b-4224-400b-b466-98e72682898e-kube-api-access-6q2s5\") pod \"crc-debug-xh575\" (UID: \"6f7e5b9b-4224-400b-b466-98e72682898e\") " pod="openshift-must-gather-2gd5d/crc-debug-xh575" Mar 21 05:57:22 crc kubenswrapper[4775]: I0321 05:57:22.326635 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q2s5\" (UniqueName: \"kubernetes.io/projected/6f7e5b9b-4224-400b-b466-98e72682898e-kube-api-access-6q2s5\") pod \"crc-debug-xh575\" (UID: \"6f7e5b9b-4224-400b-b466-98e72682898e\") " pod="openshift-must-gather-2gd5d/crc-debug-xh575" Mar 21 05:57:22 crc kubenswrapper[4775]: I0321 05:57:22.511435 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2gd5d/crc-debug-xh575" Mar 21 05:57:22 crc kubenswrapper[4775]: I0321 05:57:22.529156 4775 scope.go:117] "RemoveContainer" containerID="9b324351e03c9b33d2013ebe58b0174e6c17561e8f6660c56e5da466bb4b8c3d" Mar 21 05:57:22 crc kubenswrapper[4775]: I0321 05:57:22.529203 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2gd5d/crc-debug-8rplg" Mar 21 05:57:23 crc kubenswrapper[4775]: I0321 05:57:23.538256 4775 generic.go:334] "Generic (PLEG): container finished" podID="6f7e5b9b-4224-400b-b466-98e72682898e" containerID="695f561171dabe615374495a7ceda94563edda67471d58dbff6e220446956791" exitCode=0 Mar 21 05:57:23 crc kubenswrapper[4775]: I0321 05:57:23.538335 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2gd5d/crc-debug-xh575" event={"ID":"6f7e5b9b-4224-400b-b466-98e72682898e","Type":"ContainerDied","Data":"695f561171dabe615374495a7ceda94563edda67471d58dbff6e220446956791"} Mar 21 05:57:23 crc kubenswrapper[4775]: I0321 05:57:23.538926 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2gd5d/crc-debug-xh575" event={"ID":"6f7e5b9b-4224-400b-b466-98e72682898e","Type":"ContainerStarted","Data":"0433825eda899e4fbbddf6ea9298598eaa49439fe9100da79b0d50995098bcac"} Mar 21 05:57:23 crc kubenswrapper[4775]: I0321 05:57:23.581400 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2gd5d/crc-debug-xh575"] Mar 21 05:57:23 crc kubenswrapper[4775]: I0321 05:57:23.594196 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2gd5d/crc-debug-xh575"] Mar 21 05:57:23 crc kubenswrapper[4775]: I0321 05:57:23.674624 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90863e29-81bb-477f-846a-7f9c27ed3b89" path="/var/lib/kubelet/pods/90863e29-81bb-477f-846a-7f9c27ed3b89/volumes" Mar 21 05:57:24 crc kubenswrapper[4775]: I0321 05:57:24.651950 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2gd5d/crc-debug-xh575" Mar 21 05:57:24 crc kubenswrapper[4775]: I0321 05:57:24.755682 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6f7e5b9b-4224-400b-b466-98e72682898e-host\") pod \"6f7e5b9b-4224-400b-b466-98e72682898e\" (UID: \"6f7e5b9b-4224-400b-b466-98e72682898e\") " Mar 21 05:57:24 crc kubenswrapper[4775]: I0321 05:57:24.755806 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6q2s5\" (UniqueName: \"kubernetes.io/projected/6f7e5b9b-4224-400b-b466-98e72682898e-kube-api-access-6q2s5\") pod \"6f7e5b9b-4224-400b-b466-98e72682898e\" (UID: \"6f7e5b9b-4224-400b-b466-98e72682898e\") " Mar 21 05:57:24 crc kubenswrapper[4775]: I0321 05:57:24.755824 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6f7e5b9b-4224-400b-b466-98e72682898e-host" (OuterVolumeSpecName: "host") pod "6f7e5b9b-4224-400b-b466-98e72682898e" (UID: "6f7e5b9b-4224-400b-b466-98e72682898e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 05:57:24 crc kubenswrapper[4775]: I0321 05:57:24.756814 4775 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6f7e5b9b-4224-400b-b466-98e72682898e-host\") on node \"crc\" DevicePath \"\"" Mar 21 05:57:24 crc kubenswrapper[4775]: I0321 05:57:24.776972 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f7e5b9b-4224-400b-b466-98e72682898e-kube-api-access-6q2s5" (OuterVolumeSpecName: "kube-api-access-6q2s5") pod "6f7e5b9b-4224-400b-b466-98e72682898e" (UID: "6f7e5b9b-4224-400b-b466-98e72682898e"). InnerVolumeSpecName "kube-api-access-6q2s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:57:24 crc kubenswrapper[4775]: I0321 05:57:24.858779 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6q2s5\" (UniqueName: \"kubernetes.io/projected/6f7e5b9b-4224-400b-b466-98e72682898e-kube-api-access-6q2s5\") on node \"crc\" DevicePath \"\"" Mar 21 05:57:25 crc kubenswrapper[4775]: I0321 05:57:25.557674 4775 scope.go:117] "RemoveContainer" containerID="695f561171dabe615374495a7ceda94563edda67471d58dbff6e220446956791" Mar 21 05:57:25 crc kubenswrapper[4775]: I0321 05:57:25.557722 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2gd5d/crc-debug-xh575" Mar 21 05:57:25 crc kubenswrapper[4775]: I0321 05:57:25.671156 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f7e5b9b-4224-400b-b466-98e72682898e" path="/var/lib/kubelet/pods/6f7e5b9b-4224-400b-b466-98e72682898e/volumes" Mar 21 05:57:32 crc kubenswrapper[4775]: I0321 05:57:32.482191 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:57:32 crc kubenswrapper[4775]: I0321 05:57:32.482832 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:57:32 crc kubenswrapper[4775]: I0321 05:57:32.482885 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 05:57:32 crc kubenswrapper[4775]: I0321 05:57:32.483804 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fb650740af569c16423f24bd3abda67b78d6d33afa7ae7aa33a8a75f71e85f1c"} pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 05:57:32 crc kubenswrapper[4775]: I0321 05:57:32.483865 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" containerID="cri-o://fb650740af569c16423f24bd3abda67b78d6d33afa7ae7aa33a8a75f71e85f1c" gracePeriod=600 Mar 21 05:57:32 crc kubenswrapper[4775]: I0321 05:57:32.643909 4775 generic.go:334] "Generic (PLEG): container finished" podID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerID="fb650740af569c16423f24bd3abda67b78d6d33afa7ae7aa33a8a75f71e85f1c" exitCode=0 Mar 21 05:57:32 crc kubenswrapper[4775]: I0321 05:57:32.643974 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerDied","Data":"fb650740af569c16423f24bd3abda67b78d6d33afa7ae7aa33a8a75f71e85f1c"} Mar 21 05:57:32 crc kubenswrapper[4775]: I0321 05:57:32.644020 4775 scope.go:117] "RemoveContainer" containerID="fb52ad12aca7a430a7e7f10e8a7f0c10de8bb4e4ef886aa1ca0cb8ffad7427f2" Mar 21 05:57:33 crc kubenswrapper[4775]: I0321 05:57:33.653866 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerStarted","Data":"9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226"} Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.038425 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-77fd86567d-mf2wb_1727040d-36f5-431c-b8f1-84e206146dcf/barbican-api/0.log" Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.156672 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567878-jj8nx"] Mar 21 05:58:00 crc kubenswrapper[4775]: E0321 05:58:00.157318 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f7e5b9b-4224-400b-b466-98e72682898e" containerName="container-00" Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.157341 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f7e5b9b-4224-400b-b466-98e72682898e" containerName="container-00" Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.157572 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f7e5b9b-4224-400b-b466-98e72682898e" containerName="container-00" Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.158381 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567878-jj8nx" Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.159928 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-77fd86567d-mf2wb_1727040d-36f5-431c-b8f1-84e206146dcf/barbican-api-log/0.log" Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.169911 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567878-jj8nx"] Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.171266 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.171517 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.171903 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.251343 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5btc\" (UniqueName: \"kubernetes.io/projected/7ad9f411-868c-4132-a586-c0f66e18edbc-kube-api-access-j5btc\") pod \"auto-csr-approver-29567878-jj8nx\" (UID: \"7ad9f411-868c-4132-a586-c0f66e18edbc\") " pod="openshift-infra/auto-csr-approver-29567878-jj8nx" Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.305724 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-75f5b547c8-mgjw5_177228f6-7f69-49c2-9942-ea0a98b56d13/barbican-keystone-listener/0.log" Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.353543 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5btc\" (UniqueName: \"kubernetes.io/projected/7ad9f411-868c-4132-a586-c0f66e18edbc-kube-api-access-j5btc\") pod \"auto-csr-approver-29567878-jj8nx\" (UID: \"7ad9f411-868c-4132-a586-c0f66e18edbc\") " pod="openshift-infra/auto-csr-approver-29567878-jj8nx" Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.372680 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-75f5b547c8-mgjw5_177228f6-7f69-49c2-9942-ea0a98b56d13/barbican-keystone-listener-log/0.log" Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.376197 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5btc\" (UniqueName: \"kubernetes.io/projected/7ad9f411-868c-4132-a586-c0f66e18edbc-kube-api-access-j5btc\") pod \"auto-csr-approver-29567878-jj8nx\" (UID: \"7ad9f411-868c-4132-a586-c0f66e18edbc\") " pod="openshift-infra/auto-csr-approver-29567878-jj8nx" Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.477408 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6dd9f89d55-fdf8c_d8a7c2e5-3643-4675-9888-3c310e4f9ad4/barbican-worker/0.log" Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.491543 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567878-jj8nx" Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.535375 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6dd9f89d55-fdf8c_d8a7c2e5-3643-4675-9888-3c310e4f9ad4/barbican-worker-log/0.log" Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.852635 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-z5v8m_203df932-0574-4098-b897-ba50813f2ec1/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.859942 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_15d97495-428d-47e0-a115-99c7fd08850a/ceilometer-central-agent/0.log" Mar 21 05:58:00 crc kubenswrapper[4775]: I0321 05:58:00.989806 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_15d97495-428d-47e0-a115-99c7fd08850a/ceilometer-notification-agent/0.log" Mar 21 05:58:01 crc kubenswrapper[4775]: I0321 05:58:01.014650 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_15d97495-428d-47e0-a115-99c7fd08850a/proxy-httpd/0.log" Mar 21 05:58:01 crc kubenswrapper[4775]: I0321 05:58:01.038649 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567878-jj8nx"] Mar 21 05:58:01 crc kubenswrapper[4775]: I0321 05:58:01.112255 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_15d97495-428d-47e0-a115-99c7fd08850a/sg-core/0.log" Mar 21 05:58:01 crc kubenswrapper[4775]: I0321 05:58:01.261847 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c16d835a-1ec2-473d-b2d8-c8e7c978e140/cinder-api/0.log" Mar 21 05:58:01 crc kubenswrapper[4775]: I0321 05:58:01.285874 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c16d835a-1ec2-473d-b2d8-c8e7c978e140/cinder-api-log/0.log" Mar 21 05:58:01 crc kubenswrapper[4775]: I0321 05:58:01.549032 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d02bb319-292e-449c-8e5b-42c6859f1529/probe/0.log" Mar 21 05:58:01 crc kubenswrapper[4775]: I0321 05:58:01.551365 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d02bb319-292e-449c-8e5b-42c6859f1529/cinder-scheduler/0.log" Mar 21 05:58:01 crc kubenswrapper[4775]: I0321 05:58:01.738952 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-2d2z7_5efe4255-484c-47d7-800a-4d0dbc5cecd9/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:58:01 crc kubenswrapper[4775]: I0321 05:58:01.901101 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-ddjjn_8d4d2c78-9cdd-4623-9fdb-e277ce4d85bd/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:58:01 crc kubenswrapper[4775]: I0321 05:58:01.941067 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567878-jj8nx" event={"ID":"7ad9f411-868c-4132-a586-c0f66e18edbc","Type":"ContainerStarted","Data":"1fc6d0c3fd1580c0dedc00dd7614c3ddb4d963fc900eb40dcdffbf67947f04b0"} Mar 21 05:58:01 crc kubenswrapper[4775]: I0321 05:58:01.988754 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-mmgp9_4918607c-6074-4fb3-a0a0-8def479058a0/init/0.log" Mar 21 05:58:02 crc kubenswrapper[4775]: I0321 05:58:02.168606 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-mmgp9_4918607c-6074-4fb3-a0a0-8def479058a0/init/0.log" Mar 21 05:58:02 crc kubenswrapper[4775]: I0321 05:58:02.246502 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-mmgp9_4918607c-6074-4fb3-a0a0-8def479058a0/dnsmasq-dns/0.log" Mar 21 05:58:02 crc kubenswrapper[4775]: I0321 05:58:02.357080 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-h87nr_28040d61-c9ea-4a55-b113-db871dff679c/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:58:02 crc kubenswrapper[4775]: I0321 05:58:02.500315 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_01638b90-5e17-43b3-a3b5-90726b26e243/glance-httpd/0.log" Mar 21 05:58:02 crc kubenswrapper[4775]: I0321 05:58:02.545395 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_01638b90-5e17-43b3-a3b5-90726b26e243/glance-log/0.log" Mar 21 05:58:02 crc kubenswrapper[4775]: I0321 05:58:02.678675 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_e9e72c4b-a3fb-41eb-974a-74d24d6cdac9/glance-httpd/0.log" Mar 21 05:58:02 crc kubenswrapper[4775]: I0321 05:58:02.712172 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_e9e72c4b-a3fb-41eb-974a-74d24d6cdac9/glance-log/0.log" Mar 21 05:58:03 crc kubenswrapper[4775]: I0321 05:58:03.049429 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6496ddbdd4-v5mc5_fc6e433f-9e70-4b09-9780-403634bbe0dc/horizon/0.log" Mar 21 05:58:03 crc kubenswrapper[4775]: I0321 05:58:03.199267 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-jmx8f_8390751b-3911-4a24-a1a2-c3d1d10da875/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:58:03 crc kubenswrapper[4775]: I0321 05:58:03.317012 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6496ddbdd4-v5mc5_fc6e433f-9e70-4b09-9780-403634bbe0dc/horizon-log/0.log" Mar 21 05:58:03 crc kubenswrapper[4775]: I0321 05:58:03.613689 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-kfrz2_dbd64e65-be8d-42e7-a686-d5454932156d/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:58:03 crc kubenswrapper[4775]: I0321 05:58:03.752489 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_7ee3b7a0-9eb3-4702-8fb7-3286df60b21b/kube-state-metrics/0.log" Mar 21 05:58:03 crc kubenswrapper[4775]: I0321 05:58:03.896256 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-66c879cfdd-smnxp_9f965feb-5d82-4176-a14d-08a84c4ae794/keystone-api/0.log" Mar 21 05:58:03 crc kubenswrapper[4775]: I0321 05:58:03.965143 4775 generic.go:334] "Generic (PLEG): container finished" podID="7ad9f411-868c-4132-a586-c0f66e18edbc" containerID="a70b4f1d1db681654636f9ae7783b422ab2ef1d53af567edd94ba3f694c4a39b" exitCode=0 Mar 21 05:58:03 crc kubenswrapper[4775]: I0321 05:58:03.965190 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567878-jj8nx" event={"ID":"7ad9f411-868c-4132-a586-c0f66e18edbc","Type":"ContainerDied","Data":"a70b4f1d1db681654636f9ae7783b422ab2ef1d53af567edd94ba3f694c4a39b"} Mar 21 05:58:04 crc kubenswrapper[4775]: I0321 05:58:04.414534 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7d65998c7c-prp5b_3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef/neutron-api/0.log" Mar 21 05:58:04 crc kubenswrapper[4775]: I0321 05:58:04.432641 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7d65998c7c-prp5b_3c1e6a0a-1e89-4c30-8d9e-57306d73e4ef/neutron-httpd/0.log" Mar 21 05:58:04 crc kubenswrapper[4775]: I0321 05:58:04.706962 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-n2jxz_a71dcc90-c70a-4ff8-bf4a-42f1a2415827/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:58:04 crc kubenswrapper[4775]: I0321 05:58:04.786739 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-z7cd9_29e22bcc-ed74-4093-9cf8-f7a2e3ab9c70/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:58:05 crc kubenswrapper[4775]: I0321 05:58:05.420528 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567878-jj8nx" Mar 21 05:58:05 crc kubenswrapper[4775]: I0321 05:58:05.492666 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_789a0bb8-b131-4144-9400-7c32a604d6d5/nova-cell0-conductor-conductor/0.log" Mar 21 05:58:05 crc kubenswrapper[4775]: I0321 05:58:05.517876 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_886c404c-ceec-48e7-90da-96d6aa201152/nova-api-log/0.log" Mar 21 05:58:05 crc kubenswrapper[4775]: I0321 05:58:05.558128 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5btc\" (UniqueName: \"kubernetes.io/projected/7ad9f411-868c-4132-a586-c0f66e18edbc-kube-api-access-j5btc\") pod \"7ad9f411-868c-4132-a586-c0f66e18edbc\" (UID: \"7ad9f411-868c-4132-a586-c0f66e18edbc\") " Mar 21 05:58:05 crc kubenswrapper[4775]: I0321 05:58:05.570529 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ad9f411-868c-4132-a586-c0f66e18edbc-kube-api-access-j5btc" (OuterVolumeSpecName: "kube-api-access-j5btc") pod "7ad9f411-868c-4132-a586-c0f66e18edbc" (UID: "7ad9f411-868c-4132-a586-c0f66e18edbc"). InnerVolumeSpecName "kube-api-access-j5btc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 05:58:05 crc kubenswrapper[4775]: I0321 05:58:05.665743 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5btc\" (UniqueName: \"kubernetes.io/projected/7ad9f411-868c-4132-a586-c0f66e18edbc-kube-api-access-j5btc\") on node \"crc\" DevicePath \"\"" Mar 21 05:58:05 crc kubenswrapper[4775]: I0321 05:58:05.842707 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_b29323a7-476f-4a13-8085-4b2158a68850/nova-cell1-conductor-conductor/0.log" Mar 21 05:58:05 crc kubenswrapper[4775]: I0321 05:58:05.966541 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_3b7ea443-e30d-41d1-9f42-0bef9d7bd012/nova-cell1-novncproxy-novncproxy/0.log" Mar 21 05:58:05 crc kubenswrapper[4775]: I0321 05:58:05.985191 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567878-jj8nx" Mar 21 05:58:05 crc kubenswrapper[4775]: I0321 05:58:05.985172 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567878-jj8nx" event={"ID":"7ad9f411-868c-4132-a586-c0f66e18edbc","Type":"ContainerDied","Data":"1fc6d0c3fd1580c0dedc00dd7614c3ddb4d963fc900eb40dcdffbf67947f04b0"} Mar 21 05:58:05 crc kubenswrapper[4775]: I0321 05:58:05.985963 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fc6d0c3fd1580c0dedc00dd7614c3ddb4d963fc900eb40dcdffbf67947f04b0" Mar 21 05:58:06 crc kubenswrapper[4775]: I0321 05:58:06.181070 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_886c404c-ceec-48e7-90da-96d6aa201152/nova-api-api/0.log" Mar 21 05:58:06 crc kubenswrapper[4775]: I0321 05:58:06.506896 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567872-ctnt2"] Mar 21 05:58:06 crc kubenswrapper[4775]: I0321 05:58:06.515982 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567872-ctnt2"] Mar 21 05:58:06 crc kubenswrapper[4775]: I0321 05:58:06.533017 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_208cfa71-8242-4958-b9db-21fc180a6697/nova-metadata-log/0.log" Mar 21 05:58:07 crc kubenswrapper[4775]: I0321 05:58:07.017939 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-t85b7_50003f97-774c-4321-9ddf-6ac67546b19f/nova-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:58:07 crc kubenswrapper[4775]: I0321 05:58:07.163549 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_bab55731-40da-4831-a8b5-f9c413452367/nova-scheduler-scheduler/0.log" Mar 21 05:58:07 crc kubenswrapper[4775]: I0321 05:58:07.182726 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bade9789-f227-44ab-b7fa-2173445cd381/mysql-bootstrap/0.log" Mar 21 05:58:07 crc kubenswrapper[4775]: I0321 05:58:07.208005 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_208cfa71-8242-4958-b9db-21fc180a6697/nova-metadata-metadata/0.log" Mar 21 05:58:07 crc kubenswrapper[4775]: I0321 05:58:07.466894 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_76d205b7-bc2e-4dad-b513-457ff20d67e1/mysql-bootstrap/0.log" Mar 21 05:58:07 crc kubenswrapper[4775]: I0321 05:58:07.533194 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bade9789-f227-44ab-b7fa-2173445cd381/mysql-bootstrap/0.log" Mar 21 05:58:07 crc kubenswrapper[4775]: I0321 05:58:07.542385 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bade9789-f227-44ab-b7fa-2173445cd381/galera/0.log" Mar 21 05:58:07 crc kubenswrapper[4775]: I0321 05:58:07.683217 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b19b93e1-3fea-4999-8e31-ba5cc64c91ea" path="/var/lib/kubelet/pods/b19b93e1-3fea-4999-8e31-ba5cc64c91ea/volumes" Mar 21 05:58:07 crc kubenswrapper[4775]: I0321 05:58:07.741085 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_76d205b7-bc2e-4dad-b513-457ff20d67e1/galera/0.log" Mar 21 05:58:07 crc kubenswrapper[4775]: I0321 05:58:07.760558 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_76d205b7-bc2e-4dad-b513-457ff20d67e1/mysql-bootstrap/0.log" Mar 21 05:58:07 crc kubenswrapper[4775]: I0321 05:58:07.800020 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_bb5a0456-b5c5-433a-afde-fe38740e2310/openstackclient/0.log" Mar 21 05:58:08 crc kubenswrapper[4775]: I0321 05:58:08.622553 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-nmtjx_8a8e948c-2978-40c8-961b-1b010f7ea920/ovn-controller/0.log" Mar 21 05:58:08 crc kubenswrapper[4775]: I0321 05:58:08.667817 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-ckh8b_9f8661f4-e47a-4fcc-b55e-4c1eecd0dac8/openstack-network-exporter/0.log" Mar 21 05:58:08 crc kubenswrapper[4775]: I0321 05:58:08.884359 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-frhpj_88a367f7-4951-4e7c-889f-d147676654f8/ovsdb-server-init/0.log" Mar 21 05:58:09 crc kubenswrapper[4775]: I0321 05:58:09.083628 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-frhpj_88a367f7-4951-4e7c-889f-d147676654f8/ovs-vswitchd/0.log" Mar 21 05:58:09 crc kubenswrapper[4775]: I0321 05:58:09.148425 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-frhpj_88a367f7-4951-4e7c-889f-d147676654f8/ovsdb-server/0.log" Mar 21 05:58:09 crc kubenswrapper[4775]: I0321 05:58:09.173965 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-frhpj_88a367f7-4951-4e7c-889f-d147676654f8/ovsdb-server-init/0.log" Mar 21 05:58:09 crc kubenswrapper[4775]: I0321 05:58:09.449417 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_334a5c95-becc-4389-bb6f-50e5957cded6/openstack-network-exporter/0.log" Mar 21 05:58:09 crc kubenswrapper[4775]: I0321 05:58:09.467372 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-45kkj_3a3b7bfd-72cf-4ae1-9c2b-8d61e6af8e5d/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:58:09 crc kubenswrapper[4775]: I0321 05:58:09.491950 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_334a5c95-becc-4389-bb6f-50e5957cded6/ovn-northd/0.log" Mar 21 05:58:09 crc kubenswrapper[4775]: I0321 05:58:09.665557 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e/openstack-network-exporter/0.log" Mar 21 05:58:09 crc kubenswrapper[4775]: I0321 05:58:09.772448 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b4d6ed0c-fd91-4ed6-9a1c-0a31f78ed48e/ovsdbserver-nb/0.log" Mar 21 05:58:10 crc kubenswrapper[4775]: I0321 05:58:10.470880 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-58948d8bb4-rcw89_279dff90-9d39-418a-b5e7-00333a376d16/placement-api/0.log" Mar 21 05:58:10 crc kubenswrapper[4775]: I0321 05:58:10.481812 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_93886182-fca2-42a9-a134-2243c7c7073d/openstack-network-exporter/0.log" Mar 21 05:58:10 crc kubenswrapper[4775]: I0321 05:58:10.513226 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_93886182-fca2-42a9-a134-2243c7c7073d/ovsdbserver-sb/0.log" Mar 21 05:58:10 crc kubenswrapper[4775]: I0321 05:58:10.804429 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c95486b5-f2ad-4098-912d-6749b329824b/setup-container/0.log" Mar 21 05:58:10 crc kubenswrapper[4775]: I0321 05:58:10.810972 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-58948d8bb4-rcw89_279dff90-9d39-418a-b5e7-00333a376d16/placement-log/0.log" Mar 21 05:58:11 crc kubenswrapper[4775]: I0321 05:58:11.006470 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c95486b5-f2ad-4098-912d-6749b329824b/rabbitmq/0.log" Mar 21 05:58:11 crc kubenswrapper[4775]: I0321 05:58:11.058634 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e5e83941-a38d-4ee9-b967-1dac69c5a55b/setup-container/0.log" Mar 21 05:58:11 crc kubenswrapper[4775]: I0321 05:58:11.079375 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c95486b5-f2ad-4098-912d-6749b329824b/setup-container/0.log" Mar 21 05:58:11 crc kubenswrapper[4775]: I0321 05:58:11.350829 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e5e83941-a38d-4ee9-b967-1dac69c5a55b/setup-container/0.log" Mar 21 05:58:11 crc kubenswrapper[4775]: I0321 05:58:11.375686 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-zxxq6_034f630d-d6d6-41f0-8df6-e5db37b778f3/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:58:11 crc kubenswrapper[4775]: I0321 05:58:11.423289 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e5e83941-a38d-4ee9-b967-1dac69c5a55b/rabbitmq/0.log" Mar 21 05:58:11 crc kubenswrapper[4775]: I0321 05:58:11.626506 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-c2q8m_f41367b2-433d-48f7-af75-575be4b318fc/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:58:11 crc kubenswrapper[4775]: I0321 05:58:11.652886 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-qgpzc_7c88e417-9ede-41d8-8337-79620ceb7798/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:58:11 crc kubenswrapper[4775]: I0321 05:58:11.961335 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-wtk52_8cf3d7cc-425b-4d40-a26c-a88d2b210c0a/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:58:12 crc kubenswrapper[4775]: I0321 05:58:12.072140 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-swg9k_5f45d376-4f59-4584-a545-16d4ff066232/ssh-known-hosts-edpm-deployment/0.log" Mar 21 05:58:12 crc kubenswrapper[4775]: I0321 05:58:12.289054 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-f9b88fb79-vclnv_77432545-f22c-453a-b6a7-7c932712efa9/proxy-server/0.log" Mar 21 05:58:12 crc kubenswrapper[4775]: I0321 05:58:12.373911 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-f9b88fb79-vclnv_77432545-f22c-453a-b6a7-7c932712efa9/proxy-httpd/0.log" Mar 21 05:58:12 crc kubenswrapper[4775]: I0321 05:58:12.456178 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-kzll6_9e521c27-9d67-47bc-b6ac-74fabb543d3f/swift-ring-rebalance/0.log" Mar 21 05:58:12 crc kubenswrapper[4775]: I0321 05:58:12.574570 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/account-auditor/0.log" Mar 21 05:58:12 crc kubenswrapper[4775]: I0321 05:58:12.636293 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/account-reaper/0.log" Mar 21 05:58:12 crc kubenswrapper[4775]: I0321 05:58:12.736284 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/account-replicator/0.log" Mar 21 05:58:12 crc kubenswrapper[4775]: I0321 05:58:12.843735 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/account-server/0.log" Mar 21 05:58:12 crc kubenswrapper[4775]: I0321 05:58:12.890606 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/container-auditor/0.log" Mar 21 05:58:12 crc kubenswrapper[4775]: I0321 05:58:12.912493 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/container-replicator/0.log" Mar 21 05:58:12 crc kubenswrapper[4775]: I0321 05:58:12.961290 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/container-server/0.log" Mar 21 05:58:13 crc kubenswrapper[4775]: I0321 05:58:13.080917 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/container-updater/0.log" Mar 21 05:58:13 crc kubenswrapper[4775]: I0321 05:58:13.120709 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/object-auditor/0.log" Mar 21 05:58:13 crc kubenswrapper[4775]: I0321 05:58:13.132734 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/object-expirer/0.log" Mar 21 05:58:13 crc kubenswrapper[4775]: I0321 05:58:13.226136 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/object-replicator/0.log" Mar 21 05:58:13 crc kubenswrapper[4775]: I0321 05:58:13.406857 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/object-server/0.log" Mar 21 05:58:13 crc kubenswrapper[4775]: I0321 05:58:13.414483 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/object-updater/0.log" Mar 21 05:58:13 crc kubenswrapper[4775]: I0321 05:58:13.429059 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/rsync/0.log" Mar 21 05:58:13 crc kubenswrapper[4775]: I0321 05:58:13.503642 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8e93b938-c138-4cfc-a227-e1cd648ad59a/swift-recon-cron/0.log" Mar 21 05:58:13 crc kubenswrapper[4775]: I0321 05:58:13.748364 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_1c832898-838d-423d-8ad8-512c5ee5706c/tempest-tests-tempest-tests-runner/0.log" Mar 21 05:58:14 crc kubenswrapper[4775]: I0321 05:58:14.157148 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_2c99faa4-db71-4a05-a018-9c382f33f55e/test-operator-logs-container/0.log" Mar 21 05:58:14 crc kubenswrapper[4775]: I0321 05:58:14.215786 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-2pzqg_bccbefa9-966d-44b9-bd8f-bb566649b315/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:58:14 crc kubenswrapper[4775]: I0321 05:58:14.389214 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-mfft7_268b27f0-a217-459e-9502-7b522ca6fe2c/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 05:58:25 crc kubenswrapper[4775]: I0321 05:58:25.454365 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_2a8305a2-5178-437f-a896-314b34fa595e/memcached/0.log" Mar 21 05:58:43 crc kubenswrapper[4775]: I0321 05:58:43.010255 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt_a8157e0e-bf83-4dbd-af86-d09d14e6e1b1/util/0.log" Mar 21 05:58:43 crc kubenswrapper[4775]: I0321 05:58:43.182973 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt_a8157e0e-bf83-4dbd-af86-d09d14e6e1b1/util/0.log" Mar 21 05:58:43 crc kubenswrapper[4775]: I0321 05:58:43.205330 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt_a8157e0e-bf83-4dbd-af86-d09d14e6e1b1/pull/0.log" Mar 21 05:58:43 crc kubenswrapper[4775]: I0321 05:58:43.222885 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt_a8157e0e-bf83-4dbd-af86-d09d14e6e1b1/pull/0.log" Mar 21 05:58:43 crc kubenswrapper[4775]: I0321 05:58:43.423711 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt_a8157e0e-bf83-4dbd-af86-d09d14e6e1b1/extract/0.log" Mar 21 05:58:43 crc kubenswrapper[4775]: I0321 05:58:43.432835 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt_a8157e0e-bf83-4dbd-af86-d09d14e6e1b1/util/0.log" Mar 21 05:58:43 crc kubenswrapper[4775]: I0321 05:58:43.433729 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_68694dd73c2979a60ce733190d43e876b2bb9f24864058ec18a9c495bfh5vbt_a8157e0e-bf83-4dbd-af86-d09d14e6e1b1/pull/0.log" Mar 21 05:58:43 crc kubenswrapper[4775]: I0321 05:58:43.653147 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-59bc569d95-twhxx_94e1507b-be6c-4ecf-99e5-2bdcd2cc0cef/manager/0.log" Mar 21 05:58:43 crc kubenswrapper[4775]: I0321 05:58:43.863386 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-588d4d986b-lxvtw_87dcea67-7f65-46a6-996b-3985bf1b5171/manager/0.log" Mar 21 05:58:44 crc kubenswrapper[4775]: I0321 05:58:44.066720 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-79df6bcc97-lzvsq_80932361-6406-48dd-9e4b-4e9c27813f68/manager/0.log" Mar 21 05:58:44 crc kubenswrapper[4775]: I0321 05:58:44.167554 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-67dd5f86f5-m4gqz_0e83601c-758c-4f12-b745-bb68b0c4904f/manager/0.log" Mar 21 05:58:44 crc kubenswrapper[4775]: I0321 05:58:44.399152 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-8464cc45fb-7bdbc_01ed348d-8a8a-4717-ba0d-1944b3f1c081/manager/0.log" Mar 21 05:58:44 crc kubenswrapper[4775]: I0321 05:58:44.650358 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6f787dddc9-9g8pv_f00c8c4b-874f-45ec-8a1a-e0834b3fc252/manager/0.log" Mar 21 05:58:44 crc kubenswrapper[4775]: I0321 05:58:44.966922 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-65f65cc49c-2mgp8_0a66456f-7860-4dc1-9c1c-0db69ddcc800/manager/0.log" Mar 21 05:58:44 crc kubenswrapper[4775]: I0321 05:58:44.973530 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-768b96df4c-c7pjw_49a3c6c7-9e86-495d-8d1a-486d6bfbbbdd/manager/0.log" Mar 21 05:58:45 crc kubenswrapper[4775]: I0321 05:58:45.178173 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-55f864c847-jq88h_20c78c73-daf5-481e-a4ac-62de73b5969e/manager/0.log" Mar 21 05:58:45 crc kubenswrapper[4775]: I0321 05:58:45.216557 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-8d58dc466-dn22m_9fe71acc-7d35-4d4b-ac69-e193d3f39028/manager/0.log" Mar 21 05:58:45 crc kubenswrapper[4775]: I0321 05:58:45.390835 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67ccfc9778-hf688_ec38d53e-6fe4-41b5-8548-e49fadd9d6bf/manager/0.log" Mar 21 05:58:45 crc kubenswrapper[4775]: I0321 05:58:45.460835 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-767865f676-4tbvg_9bfe7d25-53ea-484d-a481-0ea04ee2b8a8/manager/0.log" Mar 21 05:58:45 crc kubenswrapper[4775]: I0321 05:58:45.687319 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5d488d59fb-tss7r_02b6af47-2c06-480b-a838-2d742efa1045/manager/0.log" Mar 21 05:58:45 crc kubenswrapper[4775]: I0321 05:58:45.706058 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5b9f45d989-67scw_745c79b1-1bcf-4c0f-82ee-a26cbba46d48/manager/0.log" Mar 21 05:58:45 crc kubenswrapper[4775]: I0321 05:58:45.928335 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-89d64c458-rwq59_3035739a-202f-4794-bb4f-ae2342a96441/manager/0.log" Mar 21 05:58:45 crc kubenswrapper[4775]: I0321 05:58:45.973241 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-85fcfb8fbb-q2k4k_899f7d20-7208-419e-b0f8-36c7fbf2e841/operator/0.log" Mar 21 05:58:46 crc kubenswrapper[4775]: I0321 05:58:46.169913 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-hvtkg_8c7426e8-8cec-4c84-8810-03a091d87cd9/registry-server/0.log" Mar 21 05:58:46 crc kubenswrapper[4775]: I0321 05:58:46.470387 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-884679f54-ctm9h_6eeb04ad-7251-488c-bd52-b2f14f6fb68b/manager/0.log" Mar 21 05:58:46 crc kubenswrapper[4775]: I0321 05:58:46.560614 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5784578c99-jj4pt_898b32c5-9f21-4fba-90c5-a333f36addf2/manager/0.log" Mar 21 05:58:46 crc kubenswrapper[4775]: I0321 05:58:46.774578 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-jvglf_95a44b12-e027-400d-b257-99f2012251d8/operator/0.log" Mar 21 05:58:46 crc kubenswrapper[4775]: I0321 05:58:46.920902 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-c674c5965-lqmgv_5968f1d9-f4e0-4c67-923e-2494e15c4088/manager/0.log" Mar 21 05:58:47 crc kubenswrapper[4775]: I0321 05:58:47.136067 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-d6b694c5-9s82h_d1dbd80a-0782-4035-a263-b52a90f6ee0e/manager/0.log" Mar 21 05:58:47 crc kubenswrapper[4775]: I0321 05:58:47.358477 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-l59gx_9cac78ed-6325-4649-bb05-a1518ae692e9/manager/0.log" Mar 21 05:58:47 crc kubenswrapper[4775]: I0321 05:58:47.367804 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-65746ff4dc-hg4rq_907f0cdf-2d87-4d09-97af-5591d061b4f6/manager/0.log" Mar 21 05:58:47 crc kubenswrapper[4775]: I0321 05:58:47.456822 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6c4d75f7f9-gdrc5_3c9f18bd-def6-45ff-a92b-25c6f40d6bb5/manager/0.log" Mar 21 05:58:49 crc kubenswrapper[4775]: I0321 05:58:49.898016 4775 scope.go:117] "RemoveContainer" containerID="25394576ff3d5198021128077858159917ead74d0d3b8d2389ae8fb099e54be8" Mar 21 05:59:07 crc kubenswrapper[4775]: I0321 05:59:07.896964 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-d7xpn_e5224539-6d29-4bc3-9656-4665eb287e28/control-plane-machine-set-operator/0.log" Mar 21 05:59:07 crc kubenswrapper[4775]: I0321 05:59:07.989347 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5rnj6_fe3df1e1-4c22-48df-aaea-469c864f0310/kube-rbac-proxy/0.log" Mar 21 05:59:08 crc kubenswrapper[4775]: I0321 05:59:08.066736 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5rnj6_fe3df1e1-4c22-48df-aaea-469c864f0310/machine-api-operator/0.log" Mar 21 05:59:21 crc kubenswrapper[4775]: I0321 05:59:21.477748 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-c2lsw_4fdc8b75-b0a1-4ed3-9eee-6ee726dd0fbe/cert-manager-controller/0.log" Mar 21 05:59:21 crc kubenswrapper[4775]: I0321 05:59:21.617868 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-ngckr_172b2006-3394-469a-be7f-1b66d020fd45/cert-manager-cainjector/0.log" Mar 21 05:59:21 crc kubenswrapper[4775]: I0321 05:59:21.734628 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-ms9hv_670f734f-e215-441e-9b56-7251bc7f2484/cert-manager-webhook/0.log" Mar 21 05:59:32 crc kubenswrapper[4775]: I0321 05:59:32.483583 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 05:59:32 crc kubenswrapper[4775]: I0321 05:59:32.484314 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 05:59:33 crc kubenswrapper[4775]: I0321 05:59:33.891391 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-8x7x2_b0fbab95-1c88-40a3-8ccb-58bca74c8f3c/nmstate-console-plugin/0.log" Mar 21 05:59:34 crc kubenswrapper[4775]: I0321 05:59:34.177597 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-6l74f_d03e5939-1625-4597-ad3b-9edf8e8075f5/nmstate-handler/0.log" Mar 21 05:59:34 crc kubenswrapper[4775]: I0321 05:59:34.256568 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-626kh_409422f2-717f-4f82-8dae-fe01dfda7083/kube-rbac-proxy/0.log" Mar 21 05:59:34 crc kubenswrapper[4775]: I0321 05:59:34.349381 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-626kh_409422f2-717f-4f82-8dae-fe01dfda7083/nmstate-metrics/0.log" Mar 21 05:59:34 crc kubenswrapper[4775]: I0321 05:59:34.419480 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-xljlp_f4a0a79a-5b67-44f8-9ef5-304530d5e764/nmstate-operator/0.log" Mar 21 05:59:34 crc kubenswrapper[4775]: I0321 05:59:34.545172 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-rptn7_944b76e5-c8c5-4cba-9df2-9e9b87b540a8/nmstate-webhook/0.log" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.160143 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567880-mnpzk"] Mar 21 06:00:00 crc kubenswrapper[4775]: E0321 06:00:00.161368 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ad9f411-868c-4132-a586-c0f66e18edbc" containerName="oc" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.161387 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ad9f411-868c-4132-a586-c0f66e18edbc" containerName="oc" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.163798 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ad9f411-868c-4132-a586-c0f66e18edbc" containerName="oc" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.164687 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567880-mnpzk" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.168471 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.168987 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.169021 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.176291 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567880-szvrl"] Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.177749 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567880-szvrl" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.181398 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.187612 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567880-szvrl"] Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.192334 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.199182 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567880-mnpzk"] Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.313345 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e331421-c1bf-4386-b8df-666fe188438c-secret-volume\") pod \"collect-profiles-29567880-szvrl\" (UID: \"7e331421-c1bf-4386-b8df-666fe188438c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567880-szvrl" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.313428 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77jtw\" (UniqueName: \"kubernetes.io/projected/d8b58398-7316-48e6-91d4-184a3600d431-kube-api-access-77jtw\") pod \"auto-csr-approver-29567880-mnpzk\" (UID: \"d8b58398-7316-48e6-91d4-184a3600d431\") " pod="openshift-infra/auto-csr-approver-29567880-mnpzk" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.313558 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfm6q\" (UniqueName: \"kubernetes.io/projected/7e331421-c1bf-4386-b8df-666fe188438c-kube-api-access-xfm6q\") pod \"collect-profiles-29567880-szvrl\" (UID: \"7e331421-c1bf-4386-b8df-666fe188438c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567880-szvrl" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.313591 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e331421-c1bf-4386-b8df-666fe188438c-config-volume\") pod \"collect-profiles-29567880-szvrl\" (UID: \"7e331421-c1bf-4386-b8df-666fe188438c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567880-szvrl" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.414803 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e331421-c1bf-4386-b8df-666fe188438c-config-volume\") pod \"collect-profiles-29567880-szvrl\" (UID: \"7e331421-c1bf-4386-b8df-666fe188438c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567880-szvrl" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.414907 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e331421-c1bf-4386-b8df-666fe188438c-secret-volume\") pod \"collect-profiles-29567880-szvrl\" (UID: \"7e331421-c1bf-4386-b8df-666fe188438c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567880-szvrl" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.414968 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77jtw\" (UniqueName: \"kubernetes.io/projected/d8b58398-7316-48e6-91d4-184a3600d431-kube-api-access-77jtw\") pod \"auto-csr-approver-29567880-mnpzk\" (UID: \"d8b58398-7316-48e6-91d4-184a3600d431\") " pod="openshift-infra/auto-csr-approver-29567880-mnpzk" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.415140 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfm6q\" (UniqueName: \"kubernetes.io/projected/7e331421-c1bf-4386-b8df-666fe188438c-kube-api-access-xfm6q\") pod \"collect-profiles-29567880-szvrl\" (UID: \"7e331421-c1bf-4386-b8df-666fe188438c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567880-szvrl" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.416388 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e331421-c1bf-4386-b8df-666fe188438c-config-volume\") pod \"collect-profiles-29567880-szvrl\" (UID: \"7e331421-c1bf-4386-b8df-666fe188438c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567880-szvrl" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.434160 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e331421-c1bf-4386-b8df-666fe188438c-secret-volume\") pod \"collect-profiles-29567880-szvrl\" (UID: \"7e331421-c1bf-4386-b8df-666fe188438c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567880-szvrl" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.436987 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfm6q\" (UniqueName: \"kubernetes.io/projected/7e331421-c1bf-4386-b8df-666fe188438c-kube-api-access-xfm6q\") pod \"collect-profiles-29567880-szvrl\" (UID: \"7e331421-c1bf-4386-b8df-666fe188438c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567880-szvrl" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.438942 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77jtw\" (UniqueName: \"kubernetes.io/projected/d8b58398-7316-48e6-91d4-184a3600d431-kube-api-access-77jtw\") pod \"auto-csr-approver-29567880-mnpzk\" (UID: \"d8b58398-7316-48e6-91d4-184a3600d431\") " pod="openshift-infra/auto-csr-approver-29567880-mnpzk" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.491585 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567880-mnpzk" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.511210 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567880-szvrl" Mar 21 06:00:00 crc kubenswrapper[4775]: I0321 06:00:00.979692 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567880-mnpzk"] Mar 21 06:00:01 crc kubenswrapper[4775]: W0321 06:00:01.057647 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e331421_c1bf_4386_b8df_666fe188438c.slice/crio-204ed16aa9d0e149e4e3fe9954d731f73acf13815efc498c74d76339962afded WatchSource:0}: Error finding container 204ed16aa9d0e149e4e3fe9954d731f73acf13815efc498c74d76339962afded: Status 404 returned error can't find the container with id 204ed16aa9d0e149e4e3fe9954d731f73acf13815efc498c74d76339962afded Mar 21 06:00:01 crc kubenswrapper[4775]: I0321 06:00:01.058069 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567880-szvrl"] Mar 21 06:00:01 crc kubenswrapper[4775]: I0321 06:00:01.176562 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567880-szvrl" event={"ID":"7e331421-c1bf-4386-b8df-666fe188438c","Type":"ContainerStarted","Data":"204ed16aa9d0e149e4e3fe9954d731f73acf13815efc498c74d76339962afded"} Mar 21 06:00:01 crc kubenswrapper[4775]: I0321 06:00:01.177857 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567880-mnpzk" event={"ID":"d8b58398-7316-48e6-91d4-184a3600d431","Type":"ContainerStarted","Data":"7b89cbef9afa63d81b143d494e06f233d2505f8b9a3323a0aefa30926440691a"} Mar 21 06:00:01 crc kubenswrapper[4775]: E0321 06:00:01.667231 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e331421_c1bf_4386_b8df_666fe188438c.slice/crio-conmon-1ca8dee3c1c4ec0e490473322114b11216cc0630376f7af892bdd947cce0e6a0.scope\": RecentStats: unable to find data in memory cache]" Mar 21 06:00:02 crc kubenswrapper[4775]: I0321 06:00:02.207701 4775 generic.go:334] "Generic (PLEG): container finished" podID="7e331421-c1bf-4386-b8df-666fe188438c" containerID="1ca8dee3c1c4ec0e490473322114b11216cc0630376f7af892bdd947cce0e6a0" exitCode=0 Mar 21 06:00:02 crc kubenswrapper[4775]: I0321 06:00:02.207758 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567880-szvrl" event={"ID":"7e331421-c1bf-4386-b8df-666fe188438c","Type":"ContainerDied","Data":"1ca8dee3c1c4ec0e490473322114b11216cc0630376f7af892bdd947cce0e6a0"} Mar 21 06:00:02 crc kubenswrapper[4775]: I0321 06:00:02.482473 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 06:00:02 crc kubenswrapper[4775]: I0321 06:00:02.482531 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 06:00:03 crc kubenswrapper[4775]: I0321 06:00:03.537440 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567880-szvrl" Mar 21 06:00:03 crc kubenswrapper[4775]: I0321 06:00:03.678646 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e331421-c1bf-4386-b8df-666fe188438c-config-volume\") pod \"7e331421-c1bf-4386-b8df-666fe188438c\" (UID: \"7e331421-c1bf-4386-b8df-666fe188438c\") " Mar 21 06:00:03 crc kubenswrapper[4775]: I0321 06:00:03.678863 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e331421-c1bf-4386-b8df-666fe188438c-secret-volume\") pod \"7e331421-c1bf-4386-b8df-666fe188438c\" (UID: \"7e331421-c1bf-4386-b8df-666fe188438c\") " Mar 21 06:00:03 crc kubenswrapper[4775]: I0321 06:00:03.678910 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfm6q\" (UniqueName: \"kubernetes.io/projected/7e331421-c1bf-4386-b8df-666fe188438c-kube-api-access-xfm6q\") pod \"7e331421-c1bf-4386-b8df-666fe188438c\" (UID: \"7e331421-c1bf-4386-b8df-666fe188438c\") " Mar 21 06:00:03 crc kubenswrapper[4775]: I0321 06:00:03.680804 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e331421-c1bf-4386-b8df-666fe188438c-config-volume" (OuterVolumeSpecName: "config-volume") pod "7e331421-c1bf-4386-b8df-666fe188438c" (UID: "7e331421-c1bf-4386-b8df-666fe188438c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 06:00:03 crc kubenswrapper[4775]: I0321 06:00:03.688279 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e331421-c1bf-4386-b8df-666fe188438c-kube-api-access-xfm6q" (OuterVolumeSpecName: "kube-api-access-xfm6q") pod "7e331421-c1bf-4386-b8df-666fe188438c" (UID: "7e331421-c1bf-4386-b8df-666fe188438c"). InnerVolumeSpecName "kube-api-access-xfm6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 06:00:03 crc kubenswrapper[4775]: I0321 06:00:03.688950 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e331421-c1bf-4386-b8df-666fe188438c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7e331421-c1bf-4386-b8df-666fe188438c" (UID: "7e331421-c1bf-4386-b8df-666fe188438c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 06:00:03 crc kubenswrapper[4775]: I0321 06:00:03.781275 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e331421-c1bf-4386-b8df-666fe188438c-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 21 06:00:03 crc kubenswrapper[4775]: I0321 06:00:03.781574 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfm6q\" (UniqueName: \"kubernetes.io/projected/7e331421-c1bf-4386-b8df-666fe188438c-kube-api-access-xfm6q\") on node \"crc\" DevicePath \"\"" Mar 21 06:00:03 crc kubenswrapper[4775]: I0321 06:00:03.781660 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e331421-c1bf-4386-b8df-666fe188438c-config-volume\") on node \"crc\" DevicePath \"\"" Mar 21 06:00:04 crc kubenswrapper[4775]: I0321 06:00:04.224695 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567880-szvrl" event={"ID":"7e331421-c1bf-4386-b8df-666fe188438c","Type":"ContainerDied","Data":"204ed16aa9d0e149e4e3fe9954d731f73acf13815efc498c74d76339962afded"} Mar 21 06:00:04 crc kubenswrapper[4775]: I0321 06:00:04.224990 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="204ed16aa9d0e149e4e3fe9954d731f73acf13815efc498c74d76339962afded" Mar 21 06:00:04 crc kubenswrapper[4775]: I0321 06:00:04.224756 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567880-szvrl" Mar 21 06:00:04 crc kubenswrapper[4775]: I0321 06:00:04.624520 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx"] Mar 21 06:00:04 crc kubenswrapper[4775]: I0321 06:00:04.632596 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567835-47dnx"] Mar 21 06:00:04 crc kubenswrapper[4775]: I0321 06:00:04.923009 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-qq976_ac0be1f3-95f0-40a4-9a94-c74cdaad9590/controller/0.log" Mar 21 06:00:04 crc kubenswrapper[4775]: I0321 06:00:04.960506 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-qq976_ac0be1f3-95f0-40a4-9a94-c74cdaad9590/kube-rbac-proxy/0.log" Mar 21 06:00:05 crc kubenswrapper[4775]: I0321 06:00:05.112378 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-frr-files/0.log" Mar 21 06:00:05 crc kubenswrapper[4775]: I0321 06:00:05.382699 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-frr-files/0.log" Mar 21 06:00:05 crc kubenswrapper[4775]: I0321 06:00:05.393203 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-metrics/0.log" Mar 21 06:00:05 crc kubenswrapper[4775]: I0321 06:00:05.393207 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-reloader/0.log" Mar 21 06:00:05 crc kubenswrapper[4775]: I0321 06:00:05.407526 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-reloader/0.log" Mar 21 06:00:05 crc kubenswrapper[4775]: I0321 06:00:05.555472 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-reloader/0.log" Mar 21 06:00:05 crc kubenswrapper[4775]: I0321 06:00:05.586784 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-frr-files/0.log" Mar 21 06:00:05 crc kubenswrapper[4775]: I0321 06:00:05.657766 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-metrics/0.log" Mar 21 06:00:05 crc kubenswrapper[4775]: I0321 06:00:05.675315 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bebab26-7b09-4e00-adee-fffa3c06f5ab" path="/var/lib/kubelet/pods/8bebab26-7b09-4e00-adee-fffa3c06f5ab/volumes" Mar 21 06:00:05 crc kubenswrapper[4775]: I0321 06:00:05.683925 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-metrics/0.log" Mar 21 06:00:05 crc kubenswrapper[4775]: I0321 06:00:05.889200 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-metrics/0.log" Mar 21 06:00:05 crc kubenswrapper[4775]: I0321 06:00:05.905708 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-frr-files/0.log" Mar 21 06:00:05 crc kubenswrapper[4775]: I0321 06:00:05.905913 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/cp-reloader/0.log" Mar 21 06:00:05 crc kubenswrapper[4775]: I0321 06:00:05.911939 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/controller/0.log" Mar 21 06:00:06 crc kubenswrapper[4775]: I0321 06:00:06.158307 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/kube-rbac-proxy-frr/0.log" Mar 21 06:00:06 crc kubenswrapper[4775]: I0321 06:00:06.158765 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/frr-metrics/0.log" Mar 21 06:00:06 crc kubenswrapper[4775]: I0321 06:00:06.189591 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/kube-rbac-proxy/0.log" Mar 21 06:00:06 crc kubenswrapper[4775]: I0321 06:00:06.430418 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/reloader/0.log" Mar 21 06:00:06 crc kubenswrapper[4775]: I0321 06:00:06.450698 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-7jsnx_e5808d9b-074a-4948-8283-fdfea77c63bc/frr-k8s-webhook-server/0.log" Mar 21 06:00:06 crc kubenswrapper[4775]: I0321 06:00:06.883502 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6695f56dbb-f6tqn_38d80f78-fa33-49b7-99c2-62d50d1c011b/manager/0.log" Mar 21 06:00:07 crc kubenswrapper[4775]: I0321 06:00:07.320473 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-559bfcf5c-qqsvn_8c52832d-1aee-4eac-b625-24110b985402/webhook-server/0.log" Mar 21 06:00:07 crc kubenswrapper[4775]: I0321 06:00:07.403151 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cpw6m_1d9349cc-e186-40c5-bb71-c176ff4f0a0d/kube-rbac-proxy/0.log" Mar 21 06:00:07 crc kubenswrapper[4775]: I0321 06:00:07.819307 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmmgn_530a34fb-bf82-4a2e-afdd-ec646afdebcd/frr/0.log" Mar 21 06:00:07 crc kubenswrapper[4775]: I0321 06:00:07.979966 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cpw6m_1d9349cc-e186-40c5-bb71-c176ff4f0a0d/speaker/0.log" Mar 21 06:00:11 crc kubenswrapper[4775]: E0321 06:00:11.900171 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8b58398_7316_48e6_91d4_184a3600d431.slice/crio-conmon-29077d3a4390d3dea3e0e351b0e11f03950cefc5d29477b332c73c6bd74c6c09.scope\": RecentStats: unable to find data in memory cache]" Mar 21 06:00:12 crc kubenswrapper[4775]: I0321 06:00:12.299193 4775 generic.go:334] "Generic (PLEG): container finished" podID="d8b58398-7316-48e6-91d4-184a3600d431" containerID="29077d3a4390d3dea3e0e351b0e11f03950cefc5d29477b332c73c6bd74c6c09" exitCode=0 Mar 21 06:00:12 crc kubenswrapper[4775]: I0321 06:00:12.299239 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567880-mnpzk" event={"ID":"d8b58398-7316-48e6-91d4-184a3600d431","Type":"ContainerDied","Data":"29077d3a4390d3dea3e0e351b0e11f03950cefc5d29477b332c73c6bd74c6c09"} Mar 21 06:00:13 crc kubenswrapper[4775]: I0321 06:00:13.701534 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567880-mnpzk" Mar 21 06:00:13 crc kubenswrapper[4775]: I0321 06:00:13.796898 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77jtw\" (UniqueName: \"kubernetes.io/projected/d8b58398-7316-48e6-91d4-184a3600d431-kube-api-access-77jtw\") pod \"d8b58398-7316-48e6-91d4-184a3600d431\" (UID: \"d8b58398-7316-48e6-91d4-184a3600d431\") " Mar 21 06:00:13 crc kubenswrapper[4775]: I0321 06:00:13.806894 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8b58398-7316-48e6-91d4-184a3600d431-kube-api-access-77jtw" (OuterVolumeSpecName: "kube-api-access-77jtw") pod "d8b58398-7316-48e6-91d4-184a3600d431" (UID: "d8b58398-7316-48e6-91d4-184a3600d431"). InnerVolumeSpecName "kube-api-access-77jtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 06:00:13 crc kubenswrapper[4775]: I0321 06:00:13.899146 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77jtw\" (UniqueName: \"kubernetes.io/projected/d8b58398-7316-48e6-91d4-184a3600d431-kube-api-access-77jtw\") on node \"crc\" DevicePath \"\"" Mar 21 06:00:14 crc kubenswrapper[4775]: I0321 06:00:14.316460 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567880-mnpzk" event={"ID":"d8b58398-7316-48e6-91d4-184a3600d431","Type":"ContainerDied","Data":"7b89cbef9afa63d81b143d494e06f233d2505f8b9a3323a0aefa30926440691a"} Mar 21 06:00:14 crc kubenswrapper[4775]: I0321 06:00:14.316766 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b89cbef9afa63d81b143d494e06f233d2505f8b9a3323a0aefa30926440691a" Mar 21 06:00:14 crc kubenswrapper[4775]: I0321 06:00:14.316543 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567880-mnpzk" Mar 21 06:00:14 crc kubenswrapper[4775]: I0321 06:00:14.783864 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567874-8f686"] Mar 21 06:00:14 crc kubenswrapper[4775]: I0321 06:00:14.793956 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567874-8f686"] Mar 21 06:00:15 crc kubenswrapper[4775]: I0321 06:00:15.674964 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6ebee46-23d7-4d38-946e-10f7a3238243" path="/var/lib/kubelet/pods/a6ebee46-23d7-4d38-946e-10f7a3238243/volumes" Mar 21 06:00:22 crc kubenswrapper[4775]: I0321 06:00:22.497518 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp_b2ca57eb-edac-4db7-bfa4-3198d80daf99/util/0.log" Mar 21 06:00:22 crc kubenswrapper[4775]: I0321 06:00:22.753852 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp_b2ca57eb-edac-4db7-bfa4-3198d80daf99/util/0.log" Mar 21 06:00:22 crc kubenswrapper[4775]: I0321 06:00:22.754587 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp_b2ca57eb-edac-4db7-bfa4-3198d80daf99/pull/0.log" Mar 21 06:00:22 crc kubenswrapper[4775]: I0321 06:00:22.760733 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp_b2ca57eb-edac-4db7-bfa4-3198d80daf99/pull/0.log" Mar 21 06:00:23 crc kubenswrapper[4775]: I0321 06:00:23.025219 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp_b2ca57eb-edac-4db7-bfa4-3198d80daf99/pull/0.log" Mar 21 06:00:23 crc kubenswrapper[4775]: I0321 06:00:23.034635 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp_b2ca57eb-edac-4db7-bfa4-3198d80daf99/extract/0.log" Mar 21 06:00:23 crc kubenswrapper[4775]: I0321 06:00:23.046935 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874xc6qp_b2ca57eb-edac-4db7-bfa4-3198d80daf99/util/0.log" Mar 21 06:00:23 crc kubenswrapper[4775]: I0321 06:00:23.233641 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s_6252c43a-d149-46f7-ac6c-263c34980fe2/util/0.log" Mar 21 06:00:23 crc kubenswrapper[4775]: I0321 06:00:23.394622 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s_6252c43a-d149-46f7-ac6c-263c34980fe2/pull/0.log" Mar 21 06:00:23 crc kubenswrapper[4775]: I0321 06:00:23.441574 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s_6252c43a-d149-46f7-ac6c-263c34980fe2/pull/0.log" Mar 21 06:00:23 crc kubenswrapper[4775]: I0321 06:00:23.443518 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s_6252c43a-d149-46f7-ac6c-263c34980fe2/util/0.log" Mar 21 06:00:23 crc kubenswrapper[4775]: I0321 06:00:23.619832 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s_6252c43a-d149-46f7-ac6c-263c34980fe2/extract/0.log" Mar 21 06:00:23 crc kubenswrapper[4775]: I0321 06:00:23.757009 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s_6252c43a-d149-46f7-ac6c-263c34980fe2/pull/0.log" Mar 21 06:00:23 crc kubenswrapper[4775]: I0321 06:00:23.928215 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1ff62s_6252c43a-d149-46f7-ac6c-263c34980fe2/util/0.log" Mar 21 06:00:24 crc kubenswrapper[4775]: I0321 06:00:24.030401 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77cp9_08558c65-4599-4c64-bb0f-f18f94cecdec/extract-utilities/0.log" Mar 21 06:00:24 crc kubenswrapper[4775]: I0321 06:00:24.226757 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77cp9_08558c65-4599-4c64-bb0f-f18f94cecdec/extract-content/0.log" Mar 21 06:00:24 crc kubenswrapper[4775]: I0321 06:00:24.227318 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77cp9_08558c65-4599-4c64-bb0f-f18f94cecdec/extract-utilities/0.log" Mar 21 06:00:24 crc kubenswrapper[4775]: I0321 06:00:24.232559 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77cp9_08558c65-4599-4c64-bb0f-f18f94cecdec/extract-content/0.log" Mar 21 06:00:24 crc kubenswrapper[4775]: I0321 06:00:24.479104 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77cp9_08558c65-4599-4c64-bb0f-f18f94cecdec/extract-utilities/0.log" Mar 21 06:00:24 crc kubenswrapper[4775]: I0321 06:00:24.484931 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77cp9_08558c65-4599-4c64-bb0f-f18f94cecdec/extract-content/0.log" Mar 21 06:00:24 crc kubenswrapper[4775]: I0321 06:00:24.735457 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x59zn_e8513867-3729-4d93-b8ca-45ecb69c50e6/extract-utilities/0.log" Mar 21 06:00:24 crc kubenswrapper[4775]: I0321 06:00:24.875310 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x59zn_e8513867-3729-4d93-b8ca-45ecb69c50e6/extract-content/0.log" Mar 21 06:00:24 crc kubenswrapper[4775]: I0321 06:00:24.949781 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x59zn_e8513867-3729-4d93-b8ca-45ecb69c50e6/extract-utilities/0.log" Mar 21 06:00:25 crc kubenswrapper[4775]: I0321 06:00:25.056848 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x59zn_e8513867-3729-4d93-b8ca-45ecb69c50e6/extract-content/0.log" Mar 21 06:00:25 crc kubenswrapper[4775]: I0321 06:00:25.165907 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77cp9_08558c65-4599-4c64-bb0f-f18f94cecdec/registry-server/0.log" Mar 21 06:00:25 crc kubenswrapper[4775]: I0321 06:00:25.256409 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x59zn_e8513867-3729-4d93-b8ca-45ecb69c50e6/extract-utilities/0.log" Mar 21 06:00:25 crc kubenswrapper[4775]: I0321 06:00:25.338461 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x59zn_e8513867-3729-4d93-b8ca-45ecb69c50e6/extract-content/0.log" Mar 21 06:00:25 crc kubenswrapper[4775]: I0321 06:00:25.469679 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-z7fcx_59fec450-4b61-4a15-b1b5-b47dedd649a0/marketplace-operator/0.log" Mar 21 06:00:25 crc kubenswrapper[4775]: I0321 06:00:25.740752 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h4wmk_4a9e7d52-67b3-4a38-a978-6566b1c7870a/extract-utilities/0.log" Mar 21 06:00:25 crc kubenswrapper[4775]: I0321 06:00:25.962352 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x59zn_e8513867-3729-4d93-b8ca-45ecb69c50e6/registry-server/0.log" Mar 21 06:00:25 crc kubenswrapper[4775]: I0321 06:00:25.983088 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h4wmk_4a9e7d52-67b3-4a38-a978-6566b1c7870a/extract-content/0.log" Mar 21 06:00:26 crc kubenswrapper[4775]: I0321 06:00:26.029972 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h4wmk_4a9e7d52-67b3-4a38-a978-6566b1c7870a/extract-content/0.log" Mar 21 06:00:26 crc kubenswrapper[4775]: I0321 06:00:26.051534 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h4wmk_4a9e7d52-67b3-4a38-a978-6566b1c7870a/extract-utilities/0.log" Mar 21 06:00:26 crc kubenswrapper[4775]: I0321 06:00:26.255573 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h4wmk_4a9e7d52-67b3-4a38-a978-6566b1c7870a/extract-content/0.log" Mar 21 06:00:26 crc kubenswrapper[4775]: I0321 06:00:26.268670 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h4wmk_4a9e7d52-67b3-4a38-a978-6566b1c7870a/extract-utilities/0.log" Mar 21 06:00:26 crc kubenswrapper[4775]: I0321 06:00:26.441500 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h4wmk_4a9e7d52-67b3-4a38-a978-6566b1c7870a/registry-server/0.log" Mar 21 06:00:26 crc kubenswrapper[4775]: I0321 06:00:26.994184 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gslhk_373fa4a4-80b8-4b32-a9ee-b272604d3adc/extract-utilities/0.log" Mar 21 06:00:27 crc kubenswrapper[4775]: I0321 06:00:27.153822 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gslhk_373fa4a4-80b8-4b32-a9ee-b272604d3adc/extract-content/0.log" Mar 21 06:00:27 crc kubenswrapper[4775]: I0321 06:00:27.183534 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gslhk_373fa4a4-80b8-4b32-a9ee-b272604d3adc/extract-content/0.log" Mar 21 06:00:27 crc kubenswrapper[4775]: I0321 06:00:27.188836 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gslhk_373fa4a4-80b8-4b32-a9ee-b272604d3adc/extract-utilities/0.log" Mar 21 06:00:27 crc kubenswrapper[4775]: I0321 06:00:27.341708 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gslhk_373fa4a4-80b8-4b32-a9ee-b272604d3adc/extract-utilities/0.log" Mar 21 06:00:27 crc kubenswrapper[4775]: I0321 06:00:27.408270 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gslhk_373fa4a4-80b8-4b32-a9ee-b272604d3adc/extract-content/0.log" Mar 21 06:00:27 crc kubenswrapper[4775]: I0321 06:00:27.878160 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gslhk_373fa4a4-80b8-4b32-a9ee-b272604d3adc/registry-server/0.log" Mar 21 06:00:32 crc kubenswrapper[4775]: I0321 06:00:32.482681 4775 patch_prober.go:28] interesting pod/machine-config-daemon-qc7hn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 06:00:32 crc kubenswrapper[4775]: I0321 06:00:32.483361 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 06:00:32 crc kubenswrapper[4775]: I0321 06:00:32.483414 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" Mar 21 06:00:32 crc kubenswrapper[4775]: I0321 06:00:32.484246 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226"} pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 06:00:32 crc kubenswrapper[4775]: I0321 06:00:32.484306 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerName="machine-config-daemon" containerID="cri-o://9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" gracePeriod=600 Mar 21 06:00:32 crc kubenswrapper[4775]: E0321 06:00:32.633427 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:00:33 crc kubenswrapper[4775]: I0321 06:00:33.518052 4775 generic.go:334] "Generic (PLEG): container finished" podID="cffcf487-ef41-4395-81eb-e5e6358f4a32" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" exitCode=0 Mar 21 06:00:33 crc kubenswrapper[4775]: I0321 06:00:33.518114 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerDied","Data":"9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226"} Mar 21 06:00:33 crc kubenswrapper[4775]: I0321 06:00:33.518514 4775 scope.go:117] "RemoveContainer" containerID="fb650740af569c16423f24bd3abda67b78d6d33afa7ae7aa33a8a75f71e85f1c" Mar 21 06:00:33 crc kubenswrapper[4775]: I0321 06:00:33.519615 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:00:33 crc kubenswrapper[4775]: E0321 06:00:33.520525 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:00:47 crc kubenswrapper[4775]: I0321 06:00:47.679668 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:00:47 crc kubenswrapper[4775]: E0321 06:00:47.682287 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:00:50 crc kubenswrapper[4775]: I0321 06:00:50.277895 4775 scope.go:117] "RemoveContainer" containerID="b68cd9cb06e02040464cbd016f2e9b85be0a632889c514eaecf91f324598c72a" Mar 21 06:00:50 crc kubenswrapper[4775]: I0321 06:00:50.352653 4775 scope.go:117] "RemoveContainer" containerID="c3748d6274f12563ae1762dcd4c9c668e2bf7b8aecb55d574f50d517af661d16" Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.158709 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29567881-7fbgc"] Mar 21 06:01:00 crc kubenswrapper[4775]: E0321 06:01:00.159878 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8b58398-7316-48e6-91d4-184a3600d431" containerName="oc" Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.159896 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8b58398-7316-48e6-91d4-184a3600d431" containerName="oc" Mar 21 06:01:00 crc kubenswrapper[4775]: E0321 06:01:00.159950 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e331421-c1bf-4386-b8df-666fe188438c" containerName="collect-profiles" Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.159958 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e331421-c1bf-4386-b8df-666fe188438c" containerName="collect-profiles" Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.160200 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8b58398-7316-48e6-91d4-184a3600d431" containerName="oc" Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.160223 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e331421-c1bf-4386-b8df-666fe188438c" containerName="collect-profiles" Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.161003 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29567881-7fbgc" Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.191576 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29567881-7fbgc"] Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.262848 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b00fa40d-721d-46a9-b871-d280a22f9f06-config-data\") pod \"keystone-cron-29567881-7fbgc\" (UID: \"b00fa40d-721d-46a9-b871-d280a22f9f06\") " pod="openstack/keystone-cron-29567881-7fbgc" Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.262962 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b00fa40d-721d-46a9-b871-d280a22f9f06-fernet-keys\") pod \"keystone-cron-29567881-7fbgc\" (UID: \"b00fa40d-721d-46a9-b871-d280a22f9f06\") " pod="openstack/keystone-cron-29567881-7fbgc" Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.263165 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b00fa40d-721d-46a9-b871-d280a22f9f06-combined-ca-bundle\") pod \"keystone-cron-29567881-7fbgc\" (UID: \"b00fa40d-721d-46a9-b871-d280a22f9f06\") " pod="openstack/keystone-cron-29567881-7fbgc" Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.263244 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj8wz\" (UniqueName: \"kubernetes.io/projected/b00fa40d-721d-46a9-b871-d280a22f9f06-kube-api-access-nj8wz\") pod \"keystone-cron-29567881-7fbgc\" (UID: \"b00fa40d-721d-46a9-b871-d280a22f9f06\") " pod="openstack/keystone-cron-29567881-7fbgc" Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.365023 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b00fa40d-721d-46a9-b871-d280a22f9f06-combined-ca-bundle\") pod \"keystone-cron-29567881-7fbgc\" (UID: \"b00fa40d-721d-46a9-b871-d280a22f9f06\") " pod="openstack/keystone-cron-29567881-7fbgc" Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.365992 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj8wz\" (UniqueName: \"kubernetes.io/projected/b00fa40d-721d-46a9-b871-d280a22f9f06-kube-api-access-nj8wz\") pod \"keystone-cron-29567881-7fbgc\" (UID: \"b00fa40d-721d-46a9-b871-d280a22f9f06\") " pod="openstack/keystone-cron-29567881-7fbgc" Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.366071 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b00fa40d-721d-46a9-b871-d280a22f9f06-config-data\") pod \"keystone-cron-29567881-7fbgc\" (UID: \"b00fa40d-721d-46a9-b871-d280a22f9f06\") " pod="openstack/keystone-cron-29567881-7fbgc" Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.366097 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b00fa40d-721d-46a9-b871-d280a22f9f06-fernet-keys\") pod \"keystone-cron-29567881-7fbgc\" (UID: \"b00fa40d-721d-46a9-b871-d280a22f9f06\") " pod="openstack/keystone-cron-29567881-7fbgc" Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.371666 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b00fa40d-721d-46a9-b871-d280a22f9f06-fernet-keys\") pod \"keystone-cron-29567881-7fbgc\" (UID: \"b00fa40d-721d-46a9-b871-d280a22f9f06\") " pod="openstack/keystone-cron-29567881-7fbgc" Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.371667 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b00fa40d-721d-46a9-b871-d280a22f9f06-config-data\") pod \"keystone-cron-29567881-7fbgc\" (UID: \"b00fa40d-721d-46a9-b871-d280a22f9f06\") " pod="openstack/keystone-cron-29567881-7fbgc" Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.386965 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj8wz\" (UniqueName: \"kubernetes.io/projected/b00fa40d-721d-46a9-b871-d280a22f9f06-kube-api-access-nj8wz\") pod \"keystone-cron-29567881-7fbgc\" (UID: \"b00fa40d-721d-46a9-b871-d280a22f9f06\") " pod="openstack/keystone-cron-29567881-7fbgc" Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.393805 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b00fa40d-721d-46a9-b871-d280a22f9f06-combined-ca-bundle\") pod \"keystone-cron-29567881-7fbgc\" (UID: \"b00fa40d-721d-46a9-b871-d280a22f9f06\") " pod="openstack/keystone-cron-29567881-7fbgc" Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.482900 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29567881-7fbgc" Mar 21 06:01:00 crc kubenswrapper[4775]: I0321 06:01:00.662931 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:01:00 crc kubenswrapper[4775]: E0321 06:01:00.663685 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:01:01 crc kubenswrapper[4775]: I0321 06:01:01.137150 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29567881-7fbgc"] Mar 21 06:01:01 crc kubenswrapper[4775]: I0321 06:01:01.820794 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29567881-7fbgc" event={"ID":"b00fa40d-721d-46a9-b871-d280a22f9f06","Type":"ContainerStarted","Data":"33e7299b76fea0dff44ff3c0e4495d43b700f349cc86461f84700fb6cb3a6026"} Mar 21 06:01:01 crc kubenswrapper[4775]: I0321 06:01:01.821180 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29567881-7fbgc" event={"ID":"b00fa40d-721d-46a9-b871-d280a22f9f06","Type":"ContainerStarted","Data":"9571a639ded309e03008689489bd3e624e56bd15417b47a2fcd34328d78ae2f1"} Mar 21 06:01:01 crc kubenswrapper[4775]: I0321 06:01:01.846031 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29567881-7fbgc" podStartSLOduration=1.8460081019999999 podStartE2EDuration="1.846008102s" podCreationTimestamp="2026-03-21 06:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 06:01:01.842453361 +0000 UTC m=+4414.818916985" watchObservedRunningTime="2026-03-21 06:01:01.846008102 +0000 UTC m=+4414.822471726" Mar 21 06:01:04 crc kubenswrapper[4775]: I0321 06:01:04.847634 4775 generic.go:334] "Generic (PLEG): container finished" podID="b00fa40d-721d-46a9-b871-d280a22f9f06" containerID="33e7299b76fea0dff44ff3c0e4495d43b700f349cc86461f84700fb6cb3a6026" exitCode=0 Mar 21 06:01:04 crc kubenswrapper[4775]: I0321 06:01:04.847874 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29567881-7fbgc" event={"ID":"b00fa40d-721d-46a9-b871-d280a22f9f06","Type":"ContainerDied","Data":"33e7299b76fea0dff44ff3c0e4495d43b700f349cc86461f84700fb6cb3a6026"} Mar 21 06:01:06 crc kubenswrapper[4775]: I0321 06:01:06.245422 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29567881-7fbgc" Mar 21 06:01:06 crc kubenswrapper[4775]: I0321 06:01:06.393713 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nj8wz\" (UniqueName: \"kubernetes.io/projected/b00fa40d-721d-46a9-b871-d280a22f9f06-kube-api-access-nj8wz\") pod \"b00fa40d-721d-46a9-b871-d280a22f9f06\" (UID: \"b00fa40d-721d-46a9-b871-d280a22f9f06\") " Mar 21 06:01:06 crc kubenswrapper[4775]: I0321 06:01:06.393788 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b00fa40d-721d-46a9-b871-d280a22f9f06-config-data\") pod \"b00fa40d-721d-46a9-b871-d280a22f9f06\" (UID: \"b00fa40d-721d-46a9-b871-d280a22f9f06\") " Mar 21 06:01:06 crc kubenswrapper[4775]: I0321 06:01:06.394660 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b00fa40d-721d-46a9-b871-d280a22f9f06-combined-ca-bundle\") pod \"b00fa40d-721d-46a9-b871-d280a22f9f06\" (UID: \"b00fa40d-721d-46a9-b871-d280a22f9f06\") " Mar 21 06:01:06 crc kubenswrapper[4775]: I0321 06:01:06.394783 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b00fa40d-721d-46a9-b871-d280a22f9f06-fernet-keys\") pod \"b00fa40d-721d-46a9-b871-d280a22f9f06\" (UID: \"b00fa40d-721d-46a9-b871-d280a22f9f06\") " Mar 21 06:01:06 crc kubenswrapper[4775]: I0321 06:01:06.400805 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b00fa40d-721d-46a9-b871-d280a22f9f06-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b00fa40d-721d-46a9-b871-d280a22f9f06" (UID: "b00fa40d-721d-46a9-b871-d280a22f9f06"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 06:01:06 crc kubenswrapper[4775]: I0321 06:01:06.422196 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b00fa40d-721d-46a9-b871-d280a22f9f06-kube-api-access-nj8wz" (OuterVolumeSpecName: "kube-api-access-nj8wz") pod "b00fa40d-721d-46a9-b871-d280a22f9f06" (UID: "b00fa40d-721d-46a9-b871-d280a22f9f06"). InnerVolumeSpecName "kube-api-access-nj8wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 06:01:06 crc kubenswrapper[4775]: I0321 06:01:06.426727 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b00fa40d-721d-46a9-b871-d280a22f9f06-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b00fa40d-721d-46a9-b871-d280a22f9f06" (UID: "b00fa40d-721d-46a9-b871-d280a22f9f06"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 06:01:06 crc kubenswrapper[4775]: I0321 06:01:06.455031 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b00fa40d-721d-46a9-b871-d280a22f9f06-config-data" (OuterVolumeSpecName: "config-data") pod "b00fa40d-721d-46a9-b871-d280a22f9f06" (UID: "b00fa40d-721d-46a9-b871-d280a22f9f06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 06:01:06 crc kubenswrapper[4775]: I0321 06:01:06.497650 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nj8wz\" (UniqueName: \"kubernetes.io/projected/b00fa40d-721d-46a9-b871-d280a22f9f06-kube-api-access-nj8wz\") on node \"crc\" DevicePath \"\"" Mar 21 06:01:06 crc kubenswrapper[4775]: I0321 06:01:06.497692 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b00fa40d-721d-46a9-b871-d280a22f9f06-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 06:01:06 crc kubenswrapper[4775]: I0321 06:01:06.497705 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b00fa40d-721d-46a9-b871-d280a22f9f06-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 06:01:06 crc kubenswrapper[4775]: I0321 06:01:06.497715 4775 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b00fa40d-721d-46a9-b871-d280a22f9f06-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 21 06:01:06 crc kubenswrapper[4775]: I0321 06:01:06.872341 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29567881-7fbgc" event={"ID":"b00fa40d-721d-46a9-b871-d280a22f9f06","Type":"ContainerDied","Data":"9571a639ded309e03008689489bd3e624e56bd15417b47a2fcd34328d78ae2f1"} Mar 21 06:01:06 crc kubenswrapper[4775]: I0321 06:01:06.872410 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9571a639ded309e03008689489bd3e624e56bd15417b47a2fcd34328d78ae2f1" Mar 21 06:01:06 crc kubenswrapper[4775]: I0321 06:01:06.872527 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29567881-7fbgc" Mar 21 06:01:14 crc kubenswrapper[4775]: I0321 06:01:14.662056 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:01:14 crc kubenswrapper[4775]: E0321 06:01:14.662878 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:01:28 crc kubenswrapper[4775]: I0321 06:01:28.661528 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:01:28 crc kubenswrapper[4775]: E0321 06:01:28.662483 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:01:43 crc kubenswrapper[4775]: I0321 06:01:43.662597 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:01:43 crc kubenswrapper[4775]: E0321 06:01:43.663415 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:01:56 crc kubenswrapper[4775]: I0321 06:01:56.662446 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:01:56 crc kubenswrapper[4775]: E0321 06:01:56.663120 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:02:00 crc kubenswrapper[4775]: I0321 06:02:00.153670 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567882-9lbmt"] Mar 21 06:02:00 crc kubenswrapper[4775]: E0321 06:02:00.154543 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b00fa40d-721d-46a9-b871-d280a22f9f06" containerName="keystone-cron" Mar 21 06:02:00 crc kubenswrapper[4775]: I0321 06:02:00.154560 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b00fa40d-721d-46a9-b871-d280a22f9f06" containerName="keystone-cron" Mar 21 06:02:00 crc kubenswrapper[4775]: I0321 06:02:00.154783 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b00fa40d-721d-46a9-b871-d280a22f9f06" containerName="keystone-cron" Mar 21 06:02:00 crc kubenswrapper[4775]: I0321 06:02:00.155666 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567882-9lbmt" Mar 21 06:02:00 crc kubenswrapper[4775]: I0321 06:02:00.160225 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 06:02:00 crc kubenswrapper[4775]: I0321 06:02:00.160262 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 06:02:00 crc kubenswrapper[4775]: I0321 06:02:00.162476 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 06:02:00 crc kubenswrapper[4775]: I0321 06:02:00.165275 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567882-9lbmt"] Mar 21 06:02:00 crc kubenswrapper[4775]: I0321 06:02:00.230191 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7gz9\" (UniqueName: \"kubernetes.io/projected/be3c940f-819c-4433-93d3-9e374377664b-kube-api-access-s7gz9\") pod \"auto-csr-approver-29567882-9lbmt\" (UID: \"be3c940f-819c-4433-93d3-9e374377664b\") " pod="openshift-infra/auto-csr-approver-29567882-9lbmt" Mar 21 06:02:00 crc kubenswrapper[4775]: I0321 06:02:00.332607 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7gz9\" (UniqueName: \"kubernetes.io/projected/be3c940f-819c-4433-93d3-9e374377664b-kube-api-access-s7gz9\") pod \"auto-csr-approver-29567882-9lbmt\" (UID: \"be3c940f-819c-4433-93d3-9e374377664b\") " pod="openshift-infra/auto-csr-approver-29567882-9lbmt" Mar 21 06:02:00 crc kubenswrapper[4775]: I0321 06:02:00.353770 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7gz9\" (UniqueName: \"kubernetes.io/projected/be3c940f-819c-4433-93d3-9e374377664b-kube-api-access-s7gz9\") pod \"auto-csr-approver-29567882-9lbmt\" (UID: \"be3c940f-819c-4433-93d3-9e374377664b\") " pod="openshift-infra/auto-csr-approver-29567882-9lbmt" Mar 21 06:02:00 crc kubenswrapper[4775]: I0321 06:02:00.476326 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567882-9lbmt" Mar 21 06:02:00 crc kubenswrapper[4775]: I0321 06:02:00.921049 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567882-9lbmt"] Mar 21 06:02:00 crc kubenswrapper[4775]: I0321 06:02:00.928023 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 06:02:01 crc kubenswrapper[4775]: I0321 06:02:01.516413 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567882-9lbmt" event={"ID":"be3c940f-819c-4433-93d3-9e374377664b","Type":"ContainerStarted","Data":"9196cc162f91871c5b955f0e73de9b175ceeba96a5016e153c7e7124c422f85b"} Mar 21 06:02:02 crc kubenswrapper[4775]: I0321 06:02:02.535283 4775 generic.go:334] "Generic (PLEG): container finished" podID="be3c940f-819c-4433-93d3-9e374377664b" containerID="abf324484875b628710ec29da86cbdd700864e96fb28df41c8b19dfdbc513b6e" exitCode=0 Mar 21 06:02:02 crc kubenswrapper[4775]: I0321 06:02:02.535585 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567882-9lbmt" event={"ID":"be3c940f-819c-4433-93d3-9e374377664b","Type":"ContainerDied","Data":"abf324484875b628710ec29da86cbdd700864e96fb28df41c8b19dfdbc513b6e"} Mar 21 06:02:04 crc kubenswrapper[4775]: I0321 06:02:04.401604 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567882-9lbmt" Mar 21 06:02:04 crc kubenswrapper[4775]: I0321 06:02:04.516817 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7gz9\" (UniqueName: \"kubernetes.io/projected/be3c940f-819c-4433-93d3-9e374377664b-kube-api-access-s7gz9\") pod \"be3c940f-819c-4433-93d3-9e374377664b\" (UID: \"be3c940f-819c-4433-93d3-9e374377664b\") " Mar 21 06:02:04 crc kubenswrapper[4775]: I0321 06:02:04.526991 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be3c940f-819c-4433-93d3-9e374377664b-kube-api-access-s7gz9" (OuterVolumeSpecName: "kube-api-access-s7gz9") pod "be3c940f-819c-4433-93d3-9e374377664b" (UID: "be3c940f-819c-4433-93d3-9e374377664b"). InnerVolumeSpecName "kube-api-access-s7gz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 06:02:04 crc kubenswrapper[4775]: I0321 06:02:04.556885 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567882-9lbmt" event={"ID":"be3c940f-819c-4433-93d3-9e374377664b","Type":"ContainerDied","Data":"9196cc162f91871c5b955f0e73de9b175ceeba96a5016e153c7e7124c422f85b"} Mar 21 06:02:04 crc kubenswrapper[4775]: I0321 06:02:04.556924 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9196cc162f91871c5b955f0e73de9b175ceeba96a5016e153c7e7124c422f85b" Mar 21 06:02:04 crc kubenswrapper[4775]: I0321 06:02:04.557229 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567882-9lbmt" Mar 21 06:02:04 crc kubenswrapper[4775]: I0321 06:02:04.619017 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7gz9\" (UniqueName: \"kubernetes.io/projected/be3c940f-819c-4433-93d3-9e374377664b-kube-api-access-s7gz9\") on node \"crc\" DevicePath \"\"" Mar 21 06:02:04 crc kubenswrapper[4775]: E0321 06:02:04.718247 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe3c940f_819c_4433_93d3_9e374377664b.slice\": RecentStats: unable to find data in memory cache]" Mar 21 06:02:05 crc kubenswrapper[4775]: I0321 06:02:05.472050 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567876-cl4xc"] Mar 21 06:02:05 crc kubenswrapper[4775]: I0321 06:02:05.480392 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567876-cl4xc"] Mar 21 06:02:05 crc kubenswrapper[4775]: I0321 06:02:05.673181 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77bcfe47-3b2b-4864-8659-69053e25c0a7" path="/var/lib/kubelet/pods/77bcfe47-3b2b-4864-8659-69053e25c0a7/volumes" Mar 21 06:02:12 crc kubenswrapper[4775]: I0321 06:02:12.662178 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:02:12 crc kubenswrapper[4775]: E0321 06:02:12.663338 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:02:23 crc kubenswrapper[4775]: I0321 06:02:23.750780 4775 generic.go:334] "Generic (PLEG): container finished" podID="f8e9d6cc-a02d-4747-8f7c-f439b3227b11" containerID="9e98fbe03eab0e6c4baf6973b4872261d6f0c83d76001f189bcb6eab76312381" exitCode=0 Mar 21 06:02:23 crc kubenswrapper[4775]: I0321 06:02:23.750917 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2gd5d/must-gather-fmnth" event={"ID":"f8e9d6cc-a02d-4747-8f7c-f439b3227b11","Type":"ContainerDied","Data":"9e98fbe03eab0e6c4baf6973b4872261d6f0c83d76001f189bcb6eab76312381"} Mar 21 06:02:23 crc kubenswrapper[4775]: I0321 06:02:23.753453 4775 scope.go:117] "RemoveContainer" containerID="9e98fbe03eab0e6c4baf6973b4872261d6f0c83d76001f189bcb6eab76312381" Mar 21 06:02:24 crc kubenswrapper[4775]: I0321 06:02:24.366948 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2gd5d_must-gather-fmnth_f8e9d6cc-a02d-4747-8f7c-f439b3227b11/gather/0.log" Mar 21 06:02:27 crc kubenswrapper[4775]: I0321 06:02:27.670881 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:02:27 crc kubenswrapper[4775]: E0321 06:02:27.673497 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:02:35 crc kubenswrapper[4775]: I0321 06:02:35.345046 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2gd5d/must-gather-fmnth"] Mar 21 06:02:35 crc kubenswrapper[4775]: I0321 06:02:35.346200 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-2gd5d/must-gather-fmnth" podUID="f8e9d6cc-a02d-4747-8f7c-f439b3227b11" containerName="copy" containerID="cri-o://7c0d546b19a6c80ff65ea248d24b3db62171d0f126c07845976d4d5ff49b7d79" gracePeriod=2 Mar 21 06:02:35 crc kubenswrapper[4775]: I0321 06:02:35.355626 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2gd5d/must-gather-fmnth"] Mar 21 06:02:35 crc kubenswrapper[4775]: E0321 06:02:35.432266 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8e9d6cc_a02d_4747_8f7c_f439b3227b11.slice/crio-7c0d546b19a6c80ff65ea248d24b3db62171d0f126c07845976d4d5ff49b7d79.scope\": RecentStats: unable to find data in memory cache]" Mar 21 06:02:35 crc kubenswrapper[4775]: I0321 06:02:35.890922 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2gd5d_must-gather-fmnth_f8e9d6cc-a02d-4747-8f7c-f439b3227b11/copy/0.log" Mar 21 06:02:35 crc kubenswrapper[4775]: I0321 06:02:35.892104 4775 generic.go:334] "Generic (PLEG): container finished" podID="f8e9d6cc-a02d-4747-8f7c-f439b3227b11" containerID="7c0d546b19a6c80ff65ea248d24b3db62171d0f126c07845976d4d5ff49b7d79" exitCode=143 Mar 21 06:02:35 crc kubenswrapper[4775]: I0321 06:02:35.892171 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c4f80f0df39580d8656dd019ad5ecda89f94b6664de1b8c9c62c10087b36fe2" Mar 21 06:02:35 crc kubenswrapper[4775]: I0321 06:02:35.928264 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2gd5d_must-gather-fmnth_f8e9d6cc-a02d-4747-8f7c-f439b3227b11/copy/0.log" Mar 21 06:02:35 crc kubenswrapper[4775]: I0321 06:02:35.928765 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2gd5d/must-gather-fmnth" Mar 21 06:02:36 crc kubenswrapper[4775]: I0321 06:02:36.015755 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5jqm\" (UniqueName: \"kubernetes.io/projected/f8e9d6cc-a02d-4747-8f7c-f439b3227b11-kube-api-access-k5jqm\") pod \"f8e9d6cc-a02d-4747-8f7c-f439b3227b11\" (UID: \"f8e9d6cc-a02d-4747-8f7c-f439b3227b11\") " Mar 21 06:02:36 crc kubenswrapper[4775]: I0321 06:02:36.015904 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f8e9d6cc-a02d-4747-8f7c-f439b3227b11-must-gather-output\") pod \"f8e9d6cc-a02d-4747-8f7c-f439b3227b11\" (UID: \"f8e9d6cc-a02d-4747-8f7c-f439b3227b11\") " Mar 21 06:02:36 crc kubenswrapper[4775]: I0321 06:02:36.020876 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8e9d6cc-a02d-4747-8f7c-f439b3227b11-kube-api-access-k5jqm" (OuterVolumeSpecName: "kube-api-access-k5jqm") pod "f8e9d6cc-a02d-4747-8f7c-f439b3227b11" (UID: "f8e9d6cc-a02d-4747-8f7c-f439b3227b11"). InnerVolumeSpecName "kube-api-access-k5jqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 06:02:36 crc kubenswrapper[4775]: I0321 06:02:36.119094 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5jqm\" (UniqueName: \"kubernetes.io/projected/f8e9d6cc-a02d-4747-8f7c-f439b3227b11-kube-api-access-k5jqm\") on node \"crc\" DevicePath \"\"" Mar 21 06:02:36 crc kubenswrapper[4775]: I0321 06:02:36.186768 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8e9d6cc-a02d-4747-8f7c-f439b3227b11-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "f8e9d6cc-a02d-4747-8f7c-f439b3227b11" (UID: "f8e9d6cc-a02d-4747-8f7c-f439b3227b11"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 06:02:36 crc kubenswrapper[4775]: I0321 06:02:36.223472 4775 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f8e9d6cc-a02d-4747-8f7c-f439b3227b11-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 21 06:02:36 crc kubenswrapper[4775]: I0321 06:02:36.900170 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2gd5d/must-gather-fmnth" Mar 21 06:02:37 crc kubenswrapper[4775]: I0321 06:02:37.673425 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8e9d6cc-a02d-4747-8f7c-f439b3227b11" path="/var/lib/kubelet/pods/f8e9d6cc-a02d-4747-8f7c-f439b3227b11/volumes" Mar 21 06:02:38 crc kubenswrapper[4775]: I0321 06:02:38.661355 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:02:38 crc kubenswrapper[4775]: E0321 06:02:38.661922 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:02:42 crc kubenswrapper[4775]: I0321 06:02:42.640106 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-84fzc"] Mar 21 06:02:42 crc kubenswrapper[4775]: E0321 06:02:42.641001 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be3c940f-819c-4433-93d3-9e374377664b" containerName="oc" Mar 21 06:02:42 crc kubenswrapper[4775]: I0321 06:02:42.641019 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="be3c940f-819c-4433-93d3-9e374377664b" containerName="oc" Mar 21 06:02:42 crc kubenswrapper[4775]: E0321 06:02:42.641035 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8e9d6cc-a02d-4747-8f7c-f439b3227b11" containerName="gather" Mar 21 06:02:42 crc kubenswrapper[4775]: I0321 06:02:42.641043 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8e9d6cc-a02d-4747-8f7c-f439b3227b11" containerName="gather" Mar 21 06:02:42 crc kubenswrapper[4775]: E0321 06:02:42.641070 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8e9d6cc-a02d-4747-8f7c-f439b3227b11" containerName="copy" Mar 21 06:02:42 crc kubenswrapper[4775]: I0321 06:02:42.641078 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8e9d6cc-a02d-4747-8f7c-f439b3227b11" containerName="copy" Mar 21 06:02:42 crc kubenswrapper[4775]: I0321 06:02:42.641340 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8e9d6cc-a02d-4747-8f7c-f439b3227b11" containerName="copy" Mar 21 06:02:42 crc kubenswrapper[4775]: I0321 06:02:42.641363 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="be3c940f-819c-4433-93d3-9e374377664b" containerName="oc" Mar 21 06:02:42 crc kubenswrapper[4775]: I0321 06:02:42.641378 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8e9d6cc-a02d-4747-8f7c-f439b3227b11" containerName="gather" Mar 21 06:02:42 crc kubenswrapper[4775]: I0321 06:02:42.642982 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-84fzc" Mar 21 06:02:42 crc kubenswrapper[4775]: I0321 06:02:42.658086 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-84fzc"] Mar 21 06:02:42 crc kubenswrapper[4775]: I0321 06:02:42.751349 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69163dff-f119-4580-8b11-b76aa49d1a3e-catalog-content\") pod \"redhat-operators-84fzc\" (UID: \"69163dff-f119-4580-8b11-b76aa49d1a3e\") " pod="openshift-marketplace/redhat-operators-84fzc" Mar 21 06:02:42 crc kubenswrapper[4775]: I0321 06:02:42.751449 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69163dff-f119-4580-8b11-b76aa49d1a3e-utilities\") pod \"redhat-operators-84fzc\" (UID: \"69163dff-f119-4580-8b11-b76aa49d1a3e\") " pod="openshift-marketplace/redhat-operators-84fzc" Mar 21 06:02:42 crc kubenswrapper[4775]: I0321 06:02:42.751553 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv4zx\" (UniqueName: \"kubernetes.io/projected/69163dff-f119-4580-8b11-b76aa49d1a3e-kube-api-access-dv4zx\") pod \"redhat-operators-84fzc\" (UID: \"69163dff-f119-4580-8b11-b76aa49d1a3e\") " pod="openshift-marketplace/redhat-operators-84fzc" Mar 21 06:02:42 crc kubenswrapper[4775]: I0321 06:02:42.853778 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69163dff-f119-4580-8b11-b76aa49d1a3e-catalog-content\") pod \"redhat-operators-84fzc\" (UID: \"69163dff-f119-4580-8b11-b76aa49d1a3e\") " pod="openshift-marketplace/redhat-operators-84fzc" Mar 21 06:02:42 crc kubenswrapper[4775]: I0321 06:02:42.853913 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69163dff-f119-4580-8b11-b76aa49d1a3e-utilities\") pod \"redhat-operators-84fzc\" (UID: \"69163dff-f119-4580-8b11-b76aa49d1a3e\") " pod="openshift-marketplace/redhat-operators-84fzc" Mar 21 06:02:42 crc kubenswrapper[4775]: I0321 06:02:42.854075 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv4zx\" (UniqueName: \"kubernetes.io/projected/69163dff-f119-4580-8b11-b76aa49d1a3e-kube-api-access-dv4zx\") pod \"redhat-operators-84fzc\" (UID: \"69163dff-f119-4580-8b11-b76aa49d1a3e\") " pod="openshift-marketplace/redhat-operators-84fzc" Mar 21 06:02:42 crc kubenswrapper[4775]: I0321 06:02:42.854395 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69163dff-f119-4580-8b11-b76aa49d1a3e-catalog-content\") pod \"redhat-operators-84fzc\" (UID: \"69163dff-f119-4580-8b11-b76aa49d1a3e\") " pod="openshift-marketplace/redhat-operators-84fzc" Mar 21 06:02:42 crc kubenswrapper[4775]: I0321 06:02:42.854636 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69163dff-f119-4580-8b11-b76aa49d1a3e-utilities\") pod \"redhat-operators-84fzc\" (UID: \"69163dff-f119-4580-8b11-b76aa49d1a3e\") " pod="openshift-marketplace/redhat-operators-84fzc" Mar 21 06:02:42 crc kubenswrapper[4775]: I0321 06:02:42.892229 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv4zx\" (UniqueName: \"kubernetes.io/projected/69163dff-f119-4580-8b11-b76aa49d1a3e-kube-api-access-dv4zx\") pod \"redhat-operators-84fzc\" (UID: \"69163dff-f119-4580-8b11-b76aa49d1a3e\") " pod="openshift-marketplace/redhat-operators-84fzc" Mar 21 06:02:42 crc kubenswrapper[4775]: I0321 06:02:42.970041 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-84fzc" Mar 21 06:02:43 crc kubenswrapper[4775]: I0321 06:02:43.466278 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-84fzc"] Mar 21 06:02:43 crc kubenswrapper[4775]: I0321 06:02:43.966225 4775 generic.go:334] "Generic (PLEG): container finished" podID="69163dff-f119-4580-8b11-b76aa49d1a3e" containerID="544948cf71c091c3b9668c3d1c57183b5e542b204afa23a4c531c026b2ee24be" exitCode=0 Mar 21 06:02:43 crc kubenswrapper[4775]: I0321 06:02:43.966286 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84fzc" event={"ID":"69163dff-f119-4580-8b11-b76aa49d1a3e","Type":"ContainerDied","Data":"544948cf71c091c3b9668c3d1c57183b5e542b204afa23a4c531c026b2ee24be"} Mar 21 06:02:43 crc kubenswrapper[4775]: I0321 06:02:43.966559 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84fzc" event={"ID":"69163dff-f119-4580-8b11-b76aa49d1a3e","Type":"ContainerStarted","Data":"9cd8085f69bacb1645282cffb1f31270cf94d93a58a88bcbb1dba70a7971766f"} Mar 21 06:02:45 crc kubenswrapper[4775]: I0321 06:02:45.991530 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84fzc" event={"ID":"69163dff-f119-4580-8b11-b76aa49d1a3e","Type":"ContainerStarted","Data":"daf99bcc88c5e92291e19ed807a395df95686881569148442d45d8d404d6c413"} Mar 21 06:02:50 crc kubenswrapper[4775]: I0321 06:02:50.030928 4775 generic.go:334] "Generic (PLEG): container finished" podID="69163dff-f119-4580-8b11-b76aa49d1a3e" containerID="daf99bcc88c5e92291e19ed807a395df95686881569148442d45d8d404d6c413" exitCode=0 Mar 21 06:02:50 crc kubenswrapper[4775]: I0321 06:02:50.030984 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84fzc" event={"ID":"69163dff-f119-4580-8b11-b76aa49d1a3e","Type":"ContainerDied","Data":"daf99bcc88c5e92291e19ed807a395df95686881569148442d45d8d404d6c413"} Mar 21 06:02:50 crc kubenswrapper[4775]: I0321 06:02:50.485481 4775 scope.go:117] "RemoveContainer" containerID="ff881789564104cd73abc465716a89e9559dbe987cab602aecfdcfd1bd6865c8" Mar 21 06:02:50 crc kubenswrapper[4775]: I0321 06:02:50.537212 4775 scope.go:117] "RemoveContainer" containerID="7c0d546b19a6c80ff65ea248d24b3db62171d0f126c07845976d4d5ff49b7d79" Mar 21 06:02:50 crc kubenswrapper[4775]: I0321 06:02:50.564597 4775 scope.go:117] "RemoveContainer" containerID="9e98fbe03eab0e6c4baf6973b4872261d6f0c83d76001f189bcb6eab76312381" Mar 21 06:02:51 crc kubenswrapper[4775]: I0321 06:02:51.044909 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84fzc" event={"ID":"69163dff-f119-4580-8b11-b76aa49d1a3e","Type":"ContainerStarted","Data":"ba6cac641ea0a1f05579a2bf5c076d00e13cb75f839708ed34b79a2988888e41"} Mar 21 06:02:52 crc kubenswrapper[4775]: I0321 06:02:52.074676 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-84fzc" podStartSLOduration=3.369285068 podStartE2EDuration="10.074654767s" podCreationTimestamp="2026-03-21 06:02:42 +0000 UTC" firstStartedPulling="2026-03-21 06:02:43.96844029 +0000 UTC m=+4516.944903914" lastFinishedPulling="2026-03-21 06:02:50.673809989 +0000 UTC m=+4523.650273613" observedRunningTime="2026-03-21 06:02:52.0705268 +0000 UTC m=+4525.046990464" watchObservedRunningTime="2026-03-21 06:02:52.074654767 +0000 UTC m=+4525.051118401" Mar 21 06:02:52 crc kubenswrapper[4775]: I0321 06:02:52.661862 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:02:52 crc kubenswrapper[4775]: E0321 06:02:52.662169 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:02:52 crc kubenswrapper[4775]: I0321 06:02:52.970668 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-84fzc" Mar 21 06:02:52 crc kubenswrapper[4775]: I0321 06:02:52.970733 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-84fzc" Mar 21 06:02:54 crc kubenswrapper[4775]: I0321 06:02:54.016739 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-84fzc" podUID="69163dff-f119-4580-8b11-b76aa49d1a3e" containerName="registry-server" probeResult="failure" output=< Mar 21 06:02:54 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Mar 21 06:02:54 crc kubenswrapper[4775]: > Mar 21 06:03:03 crc kubenswrapper[4775]: I0321 06:03:03.015967 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-84fzc" Mar 21 06:03:03 crc kubenswrapper[4775]: I0321 06:03:03.073839 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-84fzc" Mar 21 06:03:03 crc kubenswrapper[4775]: I0321 06:03:03.254904 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-84fzc"] Mar 21 06:03:04 crc kubenswrapper[4775]: I0321 06:03:04.158794 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-84fzc" podUID="69163dff-f119-4580-8b11-b76aa49d1a3e" containerName="registry-server" containerID="cri-o://ba6cac641ea0a1f05579a2bf5c076d00e13cb75f839708ed34b79a2988888e41" gracePeriod=2 Mar 21 06:03:04 crc kubenswrapper[4775]: I0321 06:03:04.667055 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-84fzc" Mar 21 06:03:04 crc kubenswrapper[4775]: I0321 06:03:04.782478 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69163dff-f119-4580-8b11-b76aa49d1a3e-utilities\") pod \"69163dff-f119-4580-8b11-b76aa49d1a3e\" (UID: \"69163dff-f119-4580-8b11-b76aa49d1a3e\") " Mar 21 06:03:04 crc kubenswrapper[4775]: I0321 06:03:04.782717 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dv4zx\" (UniqueName: \"kubernetes.io/projected/69163dff-f119-4580-8b11-b76aa49d1a3e-kube-api-access-dv4zx\") pod \"69163dff-f119-4580-8b11-b76aa49d1a3e\" (UID: \"69163dff-f119-4580-8b11-b76aa49d1a3e\") " Mar 21 06:03:04 crc kubenswrapper[4775]: I0321 06:03:04.782815 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69163dff-f119-4580-8b11-b76aa49d1a3e-catalog-content\") pod \"69163dff-f119-4580-8b11-b76aa49d1a3e\" (UID: \"69163dff-f119-4580-8b11-b76aa49d1a3e\") " Mar 21 06:03:04 crc kubenswrapper[4775]: I0321 06:03:04.783586 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69163dff-f119-4580-8b11-b76aa49d1a3e-utilities" (OuterVolumeSpecName: "utilities") pod "69163dff-f119-4580-8b11-b76aa49d1a3e" (UID: "69163dff-f119-4580-8b11-b76aa49d1a3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 06:03:04 crc kubenswrapper[4775]: I0321 06:03:04.790797 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69163dff-f119-4580-8b11-b76aa49d1a3e-kube-api-access-dv4zx" (OuterVolumeSpecName: "kube-api-access-dv4zx") pod "69163dff-f119-4580-8b11-b76aa49d1a3e" (UID: "69163dff-f119-4580-8b11-b76aa49d1a3e"). InnerVolumeSpecName "kube-api-access-dv4zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 06:03:04 crc kubenswrapper[4775]: I0321 06:03:04.884969 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69163dff-f119-4580-8b11-b76aa49d1a3e-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 06:03:04 crc kubenswrapper[4775]: I0321 06:03:04.885002 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dv4zx\" (UniqueName: \"kubernetes.io/projected/69163dff-f119-4580-8b11-b76aa49d1a3e-kube-api-access-dv4zx\") on node \"crc\" DevicePath \"\"" Mar 21 06:03:04 crc kubenswrapper[4775]: I0321 06:03:04.936930 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69163dff-f119-4580-8b11-b76aa49d1a3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69163dff-f119-4580-8b11-b76aa49d1a3e" (UID: "69163dff-f119-4580-8b11-b76aa49d1a3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 06:03:04 crc kubenswrapper[4775]: I0321 06:03:04.986801 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69163dff-f119-4580-8b11-b76aa49d1a3e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 06:03:05 crc kubenswrapper[4775]: I0321 06:03:05.169058 4775 generic.go:334] "Generic (PLEG): container finished" podID="69163dff-f119-4580-8b11-b76aa49d1a3e" containerID="ba6cac641ea0a1f05579a2bf5c076d00e13cb75f839708ed34b79a2988888e41" exitCode=0 Mar 21 06:03:05 crc kubenswrapper[4775]: I0321 06:03:05.169105 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84fzc" event={"ID":"69163dff-f119-4580-8b11-b76aa49d1a3e","Type":"ContainerDied","Data":"ba6cac641ea0a1f05579a2bf5c076d00e13cb75f839708ed34b79a2988888e41"} Mar 21 06:03:05 crc kubenswrapper[4775]: I0321 06:03:05.169154 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84fzc" event={"ID":"69163dff-f119-4580-8b11-b76aa49d1a3e","Type":"ContainerDied","Data":"9cd8085f69bacb1645282cffb1f31270cf94d93a58a88bcbb1dba70a7971766f"} Mar 21 06:03:05 crc kubenswrapper[4775]: I0321 06:03:05.169177 4775 scope.go:117] "RemoveContainer" containerID="ba6cac641ea0a1f05579a2bf5c076d00e13cb75f839708ed34b79a2988888e41" Mar 21 06:03:05 crc kubenswrapper[4775]: I0321 06:03:05.169199 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-84fzc" Mar 21 06:03:05 crc kubenswrapper[4775]: I0321 06:03:05.190343 4775 scope.go:117] "RemoveContainer" containerID="daf99bcc88c5e92291e19ed807a395df95686881569148442d45d8d404d6c413" Mar 21 06:03:05 crc kubenswrapper[4775]: I0321 06:03:05.218950 4775 scope.go:117] "RemoveContainer" containerID="544948cf71c091c3b9668c3d1c57183b5e542b204afa23a4c531c026b2ee24be" Mar 21 06:03:05 crc kubenswrapper[4775]: I0321 06:03:05.219726 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-84fzc"] Mar 21 06:03:05 crc kubenswrapper[4775]: I0321 06:03:05.230161 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-84fzc"] Mar 21 06:03:05 crc kubenswrapper[4775]: I0321 06:03:05.265392 4775 scope.go:117] "RemoveContainer" containerID="ba6cac641ea0a1f05579a2bf5c076d00e13cb75f839708ed34b79a2988888e41" Mar 21 06:03:05 crc kubenswrapper[4775]: E0321 06:03:05.266060 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba6cac641ea0a1f05579a2bf5c076d00e13cb75f839708ed34b79a2988888e41\": container with ID starting with ba6cac641ea0a1f05579a2bf5c076d00e13cb75f839708ed34b79a2988888e41 not found: ID does not exist" containerID="ba6cac641ea0a1f05579a2bf5c076d00e13cb75f839708ed34b79a2988888e41" Mar 21 06:03:05 crc kubenswrapper[4775]: I0321 06:03:05.266107 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba6cac641ea0a1f05579a2bf5c076d00e13cb75f839708ed34b79a2988888e41"} err="failed to get container status \"ba6cac641ea0a1f05579a2bf5c076d00e13cb75f839708ed34b79a2988888e41\": rpc error: code = NotFound desc = could not find container \"ba6cac641ea0a1f05579a2bf5c076d00e13cb75f839708ed34b79a2988888e41\": container with ID starting with ba6cac641ea0a1f05579a2bf5c076d00e13cb75f839708ed34b79a2988888e41 not found: ID does not exist" Mar 21 06:03:05 crc kubenswrapper[4775]: I0321 06:03:05.266158 4775 scope.go:117] "RemoveContainer" containerID="daf99bcc88c5e92291e19ed807a395df95686881569148442d45d8d404d6c413" Mar 21 06:03:05 crc kubenswrapper[4775]: E0321 06:03:05.266525 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"daf99bcc88c5e92291e19ed807a395df95686881569148442d45d8d404d6c413\": container with ID starting with daf99bcc88c5e92291e19ed807a395df95686881569148442d45d8d404d6c413 not found: ID does not exist" containerID="daf99bcc88c5e92291e19ed807a395df95686881569148442d45d8d404d6c413" Mar 21 06:03:05 crc kubenswrapper[4775]: I0321 06:03:05.266546 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daf99bcc88c5e92291e19ed807a395df95686881569148442d45d8d404d6c413"} err="failed to get container status \"daf99bcc88c5e92291e19ed807a395df95686881569148442d45d8d404d6c413\": rpc error: code = NotFound desc = could not find container \"daf99bcc88c5e92291e19ed807a395df95686881569148442d45d8d404d6c413\": container with ID starting with daf99bcc88c5e92291e19ed807a395df95686881569148442d45d8d404d6c413 not found: ID does not exist" Mar 21 06:03:05 crc kubenswrapper[4775]: I0321 06:03:05.266559 4775 scope.go:117] "RemoveContainer" containerID="544948cf71c091c3b9668c3d1c57183b5e542b204afa23a4c531c026b2ee24be" Mar 21 06:03:05 crc kubenswrapper[4775]: E0321 06:03:05.266920 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"544948cf71c091c3b9668c3d1c57183b5e542b204afa23a4c531c026b2ee24be\": container with ID starting with 544948cf71c091c3b9668c3d1c57183b5e542b204afa23a4c531c026b2ee24be not found: ID does not exist" containerID="544948cf71c091c3b9668c3d1c57183b5e542b204afa23a4c531c026b2ee24be" Mar 21 06:03:05 crc kubenswrapper[4775]: I0321 06:03:05.266941 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"544948cf71c091c3b9668c3d1c57183b5e542b204afa23a4c531c026b2ee24be"} err="failed to get container status \"544948cf71c091c3b9668c3d1c57183b5e542b204afa23a4c531c026b2ee24be\": rpc error: code = NotFound desc = could not find container \"544948cf71c091c3b9668c3d1c57183b5e542b204afa23a4c531c026b2ee24be\": container with ID starting with 544948cf71c091c3b9668c3d1c57183b5e542b204afa23a4c531c026b2ee24be not found: ID does not exist" Mar 21 06:03:05 crc kubenswrapper[4775]: I0321 06:03:05.671790 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69163dff-f119-4580-8b11-b76aa49d1a3e" path="/var/lib/kubelet/pods/69163dff-f119-4580-8b11-b76aa49d1a3e/volumes" Mar 21 06:03:07 crc kubenswrapper[4775]: I0321 06:03:07.670892 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:03:07 crc kubenswrapper[4775]: E0321 06:03:07.671524 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:03:22 crc kubenswrapper[4775]: I0321 06:03:22.661944 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:03:22 crc kubenswrapper[4775]: E0321 06:03:22.662881 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:03:35 crc kubenswrapper[4775]: I0321 06:03:35.662135 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:03:35 crc kubenswrapper[4775]: E0321 06:03:35.663003 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:03:47 crc kubenswrapper[4775]: I0321 06:03:47.692504 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:03:47 crc kubenswrapper[4775]: E0321 06:03:47.693240 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:03:58 crc kubenswrapper[4775]: I0321 06:03:58.661641 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:03:58 crc kubenswrapper[4775]: E0321 06:03:58.662502 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:04:00 crc kubenswrapper[4775]: I0321 06:04:00.138875 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567884-tsqsf"] Mar 21 06:04:00 crc kubenswrapper[4775]: E0321 06:04:00.139357 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69163dff-f119-4580-8b11-b76aa49d1a3e" containerName="registry-server" Mar 21 06:04:00 crc kubenswrapper[4775]: I0321 06:04:00.139382 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="69163dff-f119-4580-8b11-b76aa49d1a3e" containerName="registry-server" Mar 21 06:04:00 crc kubenswrapper[4775]: E0321 06:04:00.139394 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69163dff-f119-4580-8b11-b76aa49d1a3e" containerName="extract-utilities" Mar 21 06:04:00 crc kubenswrapper[4775]: I0321 06:04:00.139400 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="69163dff-f119-4580-8b11-b76aa49d1a3e" containerName="extract-utilities" Mar 21 06:04:00 crc kubenswrapper[4775]: E0321 06:04:00.139420 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69163dff-f119-4580-8b11-b76aa49d1a3e" containerName="extract-content" Mar 21 06:04:00 crc kubenswrapper[4775]: I0321 06:04:00.139427 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="69163dff-f119-4580-8b11-b76aa49d1a3e" containerName="extract-content" Mar 21 06:04:00 crc kubenswrapper[4775]: I0321 06:04:00.139629 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="69163dff-f119-4580-8b11-b76aa49d1a3e" containerName="registry-server" Mar 21 06:04:00 crc kubenswrapper[4775]: I0321 06:04:00.140316 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567884-tsqsf" Mar 21 06:04:00 crc kubenswrapper[4775]: I0321 06:04:00.142460 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 06:04:00 crc kubenswrapper[4775]: I0321 06:04:00.143401 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 06:04:00 crc kubenswrapper[4775]: I0321 06:04:00.146455 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-299gm" Mar 21 06:04:00 crc kubenswrapper[4775]: I0321 06:04:00.149701 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567884-tsqsf"] Mar 21 06:04:00 crc kubenswrapper[4775]: I0321 06:04:00.193822 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mctxv\" (UniqueName: \"kubernetes.io/projected/f42b0960-590a-4bb2-815e-6b4e25ce9dc3-kube-api-access-mctxv\") pod \"auto-csr-approver-29567884-tsqsf\" (UID: \"f42b0960-590a-4bb2-815e-6b4e25ce9dc3\") " pod="openshift-infra/auto-csr-approver-29567884-tsqsf" Mar 21 06:04:00 crc kubenswrapper[4775]: I0321 06:04:00.297783 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mctxv\" (UniqueName: \"kubernetes.io/projected/f42b0960-590a-4bb2-815e-6b4e25ce9dc3-kube-api-access-mctxv\") pod \"auto-csr-approver-29567884-tsqsf\" (UID: \"f42b0960-590a-4bb2-815e-6b4e25ce9dc3\") " pod="openshift-infra/auto-csr-approver-29567884-tsqsf" Mar 21 06:04:00 crc kubenswrapper[4775]: I0321 06:04:00.319560 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mctxv\" (UniqueName: \"kubernetes.io/projected/f42b0960-590a-4bb2-815e-6b4e25ce9dc3-kube-api-access-mctxv\") pod \"auto-csr-approver-29567884-tsqsf\" (UID: \"f42b0960-590a-4bb2-815e-6b4e25ce9dc3\") " pod="openshift-infra/auto-csr-approver-29567884-tsqsf" Mar 21 06:04:00 crc kubenswrapper[4775]: I0321 06:04:00.461957 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567884-tsqsf" Mar 21 06:04:00 crc kubenswrapper[4775]: I0321 06:04:00.965464 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567884-tsqsf"] Mar 21 06:04:00 crc kubenswrapper[4775]: W0321 06:04:00.967803 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf42b0960_590a_4bb2_815e_6b4e25ce9dc3.slice/crio-0bf0332673a1cf32952ca4af8bb8b08d38facf46b52d4741cb586d9c529963a8 WatchSource:0}: Error finding container 0bf0332673a1cf32952ca4af8bb8b08d38facf46b52d4741cb586d9c529963a8: Status 404 returned error can't find the container with id 0bf0332673a1cf32952ca4af8bb8b08d38facf46b52d4741cb586d9c529963a8 Mar 21 06:04:01 crc kubenswrapper[4775]: I0321 06:04:01.679174 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567884-tsqsf" event={"ID":"f42b0960-590a-4bb2-815e-6b4e25ce9dc3","Type":"ContainerStarted","Data":"0bf0332673a1cf32952ca4af8bb8b08d38facf46b52d4741cb586d9c529963a8"} Mar 21 06:04:02 crc kubenswrapper[4775]: I0321 06:04:02.688158 4775 generic.go:334] "Generic (PLEG): container finished" podID="f42b0960-590a-4bb2-815e-6b4e25ce9dc3" containerID="841f733e77f2e32ed72910d2f3507d038abd01a9d5957766d5755d8922550f30" exitCode=0 Mar 21 06:04:02 crc kubenswrapper[4775]: I0321 06:04:02.688206 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567884-tsqsf" event={"ID":"f42b0960-590a-4bb2-815e-6b4e25ce9dc3","Type":"ContainerDied","Data":"841f733e77f2e32ed72910d2f3507d038abd01a9d5957766d5755d8922550f30"} Mar 21 06:04:04 crc kubenswrapper[4775]: I0321 06:04:04.079676 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567884-tsqsf" Mar 21 06:04:04 crc kubenswrapper[4775]: I0321 06:04:04.179862 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mctxv\" (UniqueName: \"kubernetes.io/projected/f42b0960-590a-4bb2-815e-6b4e25ce9dc3-kube-api-access-mctxv\") pod \"f42b0960-590a-4bb2-815e-6b4e25ce9dc3\" (UID: \"f42b0960-590a-4bb2-815e-6b4e25ce9dc3\") " Mar 21 06:04:04 crc kubenswrapper[4775]: I0321 06:04:04.186001 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f42b0960-590a-4bb2-815e-6b4e25ce9dc3-kube-api-access-mctxv" (OuterVolumeSpecName: "kube-api-access-mctxv") pod "f42b0960-590a-4bb2-815e-6b4e25ce9dc3" (UID: "f42b0960-590a-4bb2-815e-6b4e25ce9dc3"). InnerVolumeSpecName "kube-api-access-mctxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 06:04:04 crc kubenswrapper[4775]: I0321 06:04:04.282167 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mctxv\" (UniqueName: \"kubernetes.io/projected/f42b0960-590a-4bb2-815e-6b4e25ce9dc3-kube-api-access-mctxv\") on node \"crc\" DevicePath \"\"" Mar 21 06:04:04 crc kubenswrapper[4775]: I0321 06:04:04.708650 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567884-tsqsf" event={"ID":"f42b0960-590a-4bb2-815e-6b4e25ce9dc3","Type":"ContainerDied","Data":"0bf0332673a1cf32952ca4af8bb8b08d38facf46b52d4741cb586d9c529963a8"} Mar 21 06:04:04 crc kubenswrapper[4775]: I0321 06:04:04.708983 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bf0332673a1cf32952ca4af8bb8b08d38facf46b52d4741cb586d9c529963a8" Mar 21 06:04:04 crc kubenswrapper[4775]: I0321 06:04:04.708715 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567884-tsqsf" Mar 21 06:04:05 crc kubenswrapper[4775]: I0321 06:04:05.147109 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567878-jj8nx"] Mar 21 06:04:05 crc kubenswrapper[4775]: I0321 06:04:05.158802 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567878-jj8nx"] Mar 21 06:04:05 crc kubenswrapper[4775]: I0321 06:04:05.676667 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ad9f411-868c-4132-a586-c0f66e18edbc" path="/var/lib/kubelet/pods/7ad9f411-868c-4132-a586-c0f66e18edbc/volumes" Mar 21 06:04:11 crc kubenswrapper[4775]: I0321 06:04:11.662100 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:04:11 crc kubenswrapper[4775]: E0321 06:04:11.663913 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:04:25 crc kubenswrapper[4775]: I0321 06:04:25.662150 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:04:25 crc kubenswrapper[4775]: E0321 06:04:25.663196 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:04:39 crc kubenswrapper[4775]: I0321 06:04:39.661853 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:04:39 crc kubenswrapper[4775]: E0321 06:04:39.662832 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:04:50 crc kubenswrapper[4775]: I0321 06:04:50.722274 4775 scope.go:117] "RemoveContainer" containerID="a70b4f1d1db681654636f9ae7783b422ab2ef1d53af567edd94ba3f694c4a39b" Mar 21 06:04:53 crc kubenswrapper[4775]: I0321 06:04:53.662319 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:04:53 crc kubenswrapper[4775]: E0321 06:04:53.663491 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:05:04 crc kubenswrapper[4775]: I0321 06:05:04.662345 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:05:04 crc kubenswrapper[4775]: E0321 06:05:04.663227 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:05:15 crc kubenswrapper[4775]: I0321 06:05:15.661238 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:05:15 crc kubenswrapper[4775]: E0321 06:05:15.661994 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:05:26 crc kubenswrapper[4775]: I0321 06:05:26.660886 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:05:26 crc kubenswrapper[4775]: E0321 06:05:26.661722 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qc7hn_openshift-machine-config-operator(cffcf487-ef41-4395-81eb-e5e6358f4a32)\"" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" podUID="cffcf487-ef41-4395-81eb-e5e6358f4a32" Mar 21 06:05:37 crc kubenswrapper[4775]: I0321 06:05:37.668315 4775 scope.go:117] "RemoveContainer" containerID="9a46f9d97e2c8b6a15fe9a04ccabbff9c60ceb48870c3b95f86f360ff5688226" Mar 21 06:05:38 crc kubenswrapper[4775]: I0321 06:05:38.572345 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qc7hn" event={"ID":"cffcf487-ef41-4395-81eb-e5e6358f4a32","Type":"ContainerStarted","Data":"9024ebf98588943f86acd69f0f23316451ade4ad0341403c68225759f21c46c1"} Mar 21 06:05:47 crc kubenswrapper[4775]: I0321 06:05:47.229486 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8zwmh"] Mar 21 06:05:47 crc kubenswrapper[4775]: E0321 06:05:47.232901 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f42b0960-590a-4bb2-815e-6b4e25ce9dc3" containerName="oc" Mar 21 06:05:47 crc kubenswrapper[4775]: I0321 06:05:47.233156 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f42b0960-590a-4bb2-815e-6b4e25ce9dc3" containerName="oc" Mar 21 06:05:47 crc kubenswrapper[4775]: I0321 06:05:47.233649 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f42b0960-590a-4bb2-815e-6b4e25ce9dc3" containerName="oc" Mar 21 06:05:47 crc kubenswrapper[4775]: I0321 06:05:47.236165 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8zwmh" Mar 21 06:05:47 crc kubenswrapper[4775]: I0321 06:05:47.253191 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8zwmh"] Mar 21 06:05:47 crc kubenswrapper[4775]: I0321 06:05:47.345865 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc56q\" (UniqueName: \"kubernetes.io/projected/9abf3c52-1211-4152-9055-2f7fab20a05c-kube-api-access-bc56q\") pod \"certified-operators-8zwmh\" (UID: \"9abf3c52-1211-4152-9055-2f7fab20a05c\") " pod="openshift-marketplace/certified-operators-8zwmh" Mar 21 06:05:47 crc kubenswrapper[4775]: I0321 06:05:47.346099 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9abf3c52-1211-4152-9055-2f7fab20a05c-catalog-content\") pod \"certified-operators-8zwmh\" (UID: \"9abf3c52-1211-4152-9055-2f7fab20a05c\") " pod="openshift-marketplace/certified-operators-8zwmh" Mar 21 06:05:47 crc kubenswrapper[4775]: I0321 06:05:47.346177 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9abf3c52-1211-4152-9055-2f7fab20a05c-utilities\") pod \"certified-operators-8zwmh\" (UID: \"9abf3c52-1211-4152-9055-2f7fab20a05c\") " pod="openshift-marketplace/certified-operators-8zwmh" Mar 21 06:05:47 crc kubenswrapper[4775]: I0321 06:05:47.448317 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc56q\" (UniqueName: \"kubernetes.io/projected/9abf3c52-1211-4152-9055-2f7fab20a05c-kube-api-access-bc56q\") pod \"certified-operators-8zwmh\" (UID: \"9abf3c52-1211-4152-9055-2f7fab20a05c\") " pod="openshift-marketplace/certified-operators-8zwmh" Mar 21 06:05:47 crc kubenswrapper[4775]: I0321 06:05:47.448418 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9abf3c52-1211-4152-9055-2f7fab20a05c-catalog-content\") pod \"certified-operators-8zwmh\" (UID: \"9abf3c52-1211-4152-9055-2f7fab20a05c\") " pod="openshift-marketplace/certified-operators-8zwmh" Mar 21 06:05:47 crc kubenswrapper[4775]: I0321 06:05:47.448465 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9abf3c52-1211-4152-9055-2f7fab20a05c-utilities\") pod \"certified-operators-8zwmh\" (UID: \"9abf3c52-1211-4152-9055-2f7fab20a05c\") " pod="openshift-marketplace/certified-operators-8zwmh" Mar 21 06:05:47 crc kubenswrapper[4775]: I0321 06:05:47.448999 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9abf3c52-1211-4152-9055-2f7fab20a05c-catalog-content\") pod \"certified-operators-8zwmh\" (UID: \"9abf3c52-1211-4152-9055-2f7fab20a05c\") " pod="openshift-marketplace/certified-operators-8zwmh" Mar 21 06:05:47 crc kubenswrapper[4775]: I0321 06:05:47.449093 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9abf3c52-1211-4152-9055-2f7fab20a05c-utilities\") pod \"certified-operators-8zwmh\" (UID: \"9abf3c52-1211-4152-9055-2f7fab20a05c\") " pod="openshift-marketplace/certified-operators-8zwmh" Mar 21 06:05:47 crc kubenswrapper[4775]: I0321 06:05:47.467785 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc56q\" (UniqueName: \"kubernetes.io/projected/9abf3c52-1211-4152-9055-2f7fab20a05c-kube-api-access-bc56q\") pod \"certified-operators-8zwmh\" (UID: \"9abf3c52-1211-4152-9055-2f7fab20a05c\") " pod="openshift-marketplace/certified-operators-8zwmh" Mar 21 06:05:47 crc kubenswrapper[4775]: I0321 06:05:47.568231 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8zwmh" Mar 21 06:05:48 crc kubenswrapper[4775]: I0321 06:05:48.105407 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8zwmh"] Mar 21 06:05:48 crc kubenswrapper[4775]: W0321 06:05:48.116263 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9abf3c52_1211_4152_9055_2f7fab20a05c.slice/crio-9510d881753680b23173127d27b1c5a3df1e24eb0470c5787f57c2dbe06568c7 WatchSource:0}: Error finding container 9510d881753680b23173127d27b1c5a3df1e24eb0470c5787f57c2dbe06568c7: Status 404 returned error can't find the container with id 9510d881753680b23173127d27b1c5a3df1e24eb0470c5787f57c2dbe06568c7 Mar 21 06:05:48 crc kubenswrapper[4775]: I0321 06:05:48.702048 4775 generic.go:334] "Generic (PLEG): container finished" podID="9abf3c52-1211-4152-9055-2f7fab20a05c" containerID="534acc5a77390359561da3780439f94a536281e282223833950dbd423176d301" exitCode=0 Mar 21 06:05:48 crc kubenswrapper[4775]: I0321 06:05:48.702099 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8zwmh" event={"ID":"9abf3c52-1211-4152-9055-2f7fab20a05c","Type":"ContainerDied","Data":"534acc5a77390359561da3780439f94a536281e282223833950dbd423176d301"} Mar 21 06:05:48 crc kubenswrapper[4775]: I0321 06:05:48.702154 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8zwmh" event={"ID":"9abf3c52-1211-4152-9055-2f7fab20a05c","Type":"ContainerStarted","Data":"9510d881753680b23173127d27b1c5a3df1e24eb0470c5787f57c2dbe06568c7"}